What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 5 questions

Less than a minute. Find out how much you really know about Google search.

🕒 ~1 min 🎯 5 questions

Official statement

Excluded pages count towards the crawl budget, but they are crawled much less frequently than valid pages. If a site is nearing its crawl limit, important pages take priority.
22:11
🎥 Source video

Extracted from a Google Search Central video

⏱ 1h14 💬 EN 📅 09/08/2019 ✂ 15 statements
Watch on YouTube (22:11) →
Other statements from this video 14
  1. 1:43 Faut-il vraiment traiter Googlebot comme un utilisateur américain ?
  2. 3:29 Faut-il modifier son domaine principal dans Search Console lors d'une redirection vers une sous-page ?
  3. 5:27 Pourquoi Google a-t-il supprimé la découverte des ressources bloquées dans Search Console ?
  4. 10:46 Faut-il éviter JavaScript pour générer ses balises meta ?
  5. 27:01 Les thèmes WordPress préfabriqués pénalisent-ils vraiment votre SEO ?
  6. 27:18 Faut-il vraiment abandonner le nofollow en maillage interne pour éviter les pages de porte ?
  7. 28:35 Le test mobile-friendly suffit-il vraiment à valider l'indexation de votre JavaScript ?
  8. 29:43 Pourquoi intégrer des images Instagram via iframe ruine-t-il leur potentiel SEO ?
  9. 36:38 Les redirections 301 en chaîne font-elles exploser votre budget de crawl ?
  10. 39:59 Les données structurées suffisent-elles pour démontrer l'expertise et la crédibilité d'une page ?
  11. 41:31 Google peut-il modifier vos titres pour y ajouter votre marque ?
  12. 44:04 Pourquoi votre site bien classé n'affiche-t-il pas de sitelinks ni de boîte de recherche ?
  13. 48:30 ccTLD ou sous-dossier géociblé : quelle architecture choisir pour votre SEO international ?
  14. 49:16 L'API de la Search Console vous ment-elle sur vos pages indexées ?
📅
Official statement from (6 years ago)
TL;DR

Google confirms that excluded pages in the index count towards the crawl budget, but they are crawled significantly less often than valid pages. For sites nearing their crawl limit, Googlebot systematically prioritizes important pages. Specifically: a site with thousands of excluded pages may experience a slowdown in crawling its strategic content, especially if its margin for maneuver is tight.

What you need to understand

What is an excluded page and why does Google still crawl it?

An excluded page is a URL that Googlebot has discovered (via the sitemap, internal linking, or backlinks) but has decided not to index. The reasons are varied: noindex tag, duplicate content, insufficient quality, canonicalization to another URL, or exclusion via robots.txt after the initial visit.

Google continues to crawl these pages — less frequently, certainly — to check if their status has changed. A page excluded today can become indexable tomorrow if you remove the noindex or improve its quality. Googlebot thus maintains a periodic monitoring, even though it remains marginal compared to valid pages.

How is the crawl budget really consumed by these pages?

The crawl budget refers to the number of pages that Googlebot can or wants to crawl on your site within a given timeframe. This volume depends on the health of the server, the site's popularity, and the freshness of the content.

Excluded pages nibble away at this budget, albeit modestly. If your site has 10,000 excluded pages and 5,000 indexed pages, Googlebot will certainly prioritize the 5,000 indexed pages, but it will still sporadically visit the 10,000 excluded pages. For a site with a tight crawl budget — typically a large e-commerce site or a media site with hundreds of thousands of URLs — this residual consumption can delay the discovery of new strategic pages.

What does Google mean by 'important pages'?

Google never precisely defines this term, but we can reasonably deduce that it refers to pages that generate organic traffic, have external backlinks, are regularly updated, or belong to priority sections of the site (category pages, key product sheets, recent articles).

Prioritization occurs through signals of popularity and freshness: a page that receives visits, links, or frequent updates will be recrawled more often. Conversely, an excluded URL without backlinks or traffic quickly falls to the bottom of the queue.

  • Excluded pages consume crawl budget, even if their visit rate remains low compared to indexed pages.
  • For a site close to its crawl limit, this consumption can slow down the discovery of strategic content.
  • Googlebot systematically prioritizes pages that generate traffic, links, and recent updates.
  • Regular cleaning of excluded pages (removal, redirection, improvement) frees up budget for priority URLs.

SEO Expert opinion

Does this statement align with on-the-ground observations?

Yes, and it is consistent with the server log analyses we have been conducting for years. On sites with several hundred thousand pages, we observe that excluded URLs receive spaced visits from Googlebot — sometimes once a month, sometimes every three months — whereas high-traffic indexed pages are visited daily, or even multiple times a day.

The problem arises particularly on platforms that automatically generate thousands of URL variations (filter facets, infinite pagination pages, parameterized duplicates). These excluded URLs accumulate in Search Console and in the logs. Even though each consumes only a tiny fraction of the budget, the cumulative effect mechanically slows down the crawl of strategic pages. I have seen e-commerce sites lose several days in indexing new product sheets due to uncontrolled inflation of excluded URLs.

What uncertainties remain regarding this statement?

Google remains vague on the exact threshold at which a site is considered 'close to its crawl limit'. No numbers, no official metrics. We know that Search Console displays crawl statistics, but it never says: 'Warning, you are at 85% of your budget.' [To verify]: there is no public data available to quantitatively measure this limit.

Another point: Mueller claims that important pages are prioritized, but does not specify how Google calculates this importance. Is it solely based on internal PageRank? Past organic traffic? Content freshness? Update frequency? Probably a mix of all these, but without official documentation, we are navigating in the dark. In practice, we observe that pages with external backlinks and recent organic traffic rise in priority — but this is empirical observation, not a hard and fast rule.

In what cases does this rule not really apply?

For small sites (fewer than 5,000 pages), the crawl budget is never an issue. Googlebot can crawl the entirety of the site within a few hours. Even if you have 2,000 excluded pages, it won't slow down the indexing of your 500 valid pages. Mueller's statement primarily concerns larger sites: media sites, marketplaces, multi-category e-commerce sites.

Another special case: sites with a very high-performing server infrastructure and high popularity (many backlinks, strong authority) are naturally allocated a generous crawl budget by Google. Even with tens of thousands of excluded pages, the remaining budget is more than enough to cover important pages. The problem mainly arises for medium-sized sites — those with between 50,000 and 500,000 pages, lacking colossal authority, and must optimize every resource.

Practical impact and recommendations

What concrete steps should be taken to free up crawl budget?

First step: identify excluded pages via Search Console (tab 'Pages' > 'Why Pages Are Not Indexed'). Sort them by reason for exclusion: noindex, robots.txt, canonicalized, duplicate content, insufficient quality. For each category, ask yourself: is this exclusion intentional and justified?

If yes — for example, login pages, cart pages, or filter facets without SEO value — keep them excluded, but ensure they do not receive superfluous internal linking. A link from the main menu to a noindex page is pure waste. If not — for example, legitimate product sheets excluded due to 'insufficient quality' — improve the content, add unique descriptions, and restart the crawl.

What mistakes should be absolutely avoided?

Never massively block entire sections in robots.txt without thinking it through. Many sites block /search/, /filter/, /page/ thinking they are saving crawl budget, but if Googlebot cannot crawl these URLs, it can’t follow the links they contain either. Result: valid pages become orphaned and are never discovered.

Another common pitfall: leaving thousands of 404 pages lingering in the XML sitemap. Googlebot will crawl them, note the error, and periodically check them again to see if they have returned. This is wasted budget for nothing. Clean the sitemap, redirect 404s to relevant content, or remove them permanently.

How can you check if your site is optimized?

Analyze your server logs over a period of at least 30 days. See which URLs Googlebot visits most often, and which ones it neglects. If you find that strategic pages (new product sheets, recent articles) are only visited once a week while stale excluded pages are visited daily, you have a prioritization problem.

Compare the volume of excluded pages in Search Console with the volume of indexed pages. If the ratio exceeds 2:1 (two excluded pages for one indexed), it’s a warning signal. On a large site, aim for a ratio of 1:1 or less. The more you reduce the number of unnecessary excluded pages, the more Googlebot can focus on what matters.

  • Audit excluded pages in Search Console and categorize them by reason for exclusion.
  • Remove or redirect 404 pages still present in the XML sitemap.
  • Reduce internal linking to noindex or canonicalized pages without SEO value.
  • Improve the quality of excluded pages for 'low-value content' if they deserve indexing.
  • Analyze your server logs to identify URLs that Googlebot visits too often or not enough.
  • Aim for a ratio of excluded pages/indexed pages lower than 1:1 on sites with more than 50,000 pages.
Optimizing the management of excluded pages requires meticulous auditing, regular log analysis, and precise technical decision-making. On complex sites, these operations can quickly become time-consuming and require solid SEO architecture expertise. If you lack time or internal resources, engaging a specialized SEO agency allows you to benefit from an accurate diagnosis and a tailor-made action plan, without tying up your teams for weeks.

❓ Frequently Asked Questions

Les pages bloquées par le robots.txt consomment-elles aussi du crawl budget ?
Non. Si une URL est bloquée dans le robots.txt, Googlebot ne la crawle pas du tout. Elle n'apparaît donc pas dans les stats de crawl et ne consomme aucun budget. En revanche, elle peut quand même être indexée si elle reçoit des backlinks externes — Google indexe alors l'URL sans en connaître le contenu.
Combien de fois par mois Googlebot visite-t-il une page exclue en moyenne ?
Ça dépend de la popularité de la page. Une page exclue sans backlink ni trafic peut n'être visitée qu'une fois tous les trois mois, voire moins. Une page exclue mais liée depuis des pages populaires sera visitée plus souvent, parfois une fois par semaine.
Faut-il supprimer toutes les pages exclues de la Search Console ?
Non, pas systématiquement. Certaines exclusions sont légitimes : pages de connexion, paniers, facettes de filtres sans valeur SEO. L'objectif est de réduire les exclusions involontaires (pages valides mal configurées) et de limiter le maillage interne vers les pages exclues volontairement.
Un sitemap XML trop volumineux ralentit-il le crawl ?
Pas directement, mais si votre sitemap contient des milliers d'URLs exclues, 404 ou redirigées, Googlebot va les crawler inutilement. Mieux vaut un sitemap propre avec uniquement les URLs indexables et stratégiques.
Comment savoir si mon site est proche de sa limite de crawl budget ?
Analysez vos logs serveur : si Googlebot ne visite pas vos nouvelles pages stratégiques dans les 48-72h suivant leur publication, ou si le délai d'indexation s'allonge progressivement, c'est un signe que votre budget est tendu. La Search Console ne fournit malheureusement pas de métrique directe sur cette limite.
🏷 Related Topics
Domain Age & History Crawl & Indexing AI & SEO

🎥 From the same video 14

Other SEO insights extracted from this same Google Search Central video · duration 1h14 · published on 09/08/2019

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.