Official statement
Other statements from this video 8 ▾
- 5:31 Un HTML correct améliore-t-il vraiment votre classement SEO ?
- 9:17 Les canonicals suffisent-ils vraiment à gérer les doublons sans pénalité SEO ?
- 25:47 La balise noindex bloque-t-elle vraiment l'indexation de vos pages stratégiques ?
- 31:36 Les signaux sociaux influencent-ils vraiment le classement dans Google ?
- 34:19 Le PageRank influence-t-il encore vraiment le classement Google en SEO ?
- 39:58 L'achat de liens et les échanges de backlinks conduisent-ils vraiment à des pénalités ?
- 55:24 Les pages AMP exclues de l'index signalent-elles vraiment une mauvaise implémentation ?
- 67:02 Le contenu de qualité suffit-il vraiment à bien se positionner dans Google ?
Google states that pages marked as excluded in the Search Console have no impact on your site's PageRank. Only accidental exclusions—via noindex or robots.txt—are problematic and block indexing. For SEO, this means it's crucial to distinguish between intentional exclusions and technical errors that deprive strategic pages of their ability to pass link equity.
What you need to understand
What does 'excluded' really mean in the Search Console?
The Search Console groups under 'excluded' all the pages that Google has discovered but does not index. This catch-all category mixes very different situations: pages blocked by robots.txt, marked noindex, detected as duplicates, or simply ignored because deemed worthless.
The problem is that this vague terminology creates confusion. An 'excluded' page can be a page that you want to exclude—such as a thank-you page or a facet filter—or a page you wanted to index but that Google has disregarded. The distinction is not trivial.
Why does Google specify that these pages do not affect PageRank?
Google is responding here to a widespread belief: that thousands of pages excluded in the Search Console would dilute internal PageRank or penalize the site. This fear is unfounded if the exclusions are intentional and legitimate.
On the other hand, if strategic pages are accidentally excluded—due to a forgotten noindex or a poorly calibrated robots.txt block—they cannot be indexed or pass their equity. Here, the SEO impact becomes real, even if it's not the number of excluded pages that is the issue, but their nature.
How can you distinguish between an intentional exclusion and a technical error?
An intentional exclusion results from a directive that you put in place: noindex tag, canonical pointing to another URL, targeted robots.txt block. An accidental exclusion is when a page you want to index ends up on this list without your intent.
In practical terms, a thorough audit of the Search Console is necessary. Cross-reference the list of excluded URLs with your crawl plan and editorial priorities. If high-potential pages—product sheets, landing pages—appear as excluded, it’s a warning signal.
- Intentionally excluded pages (filters, technical pages) have no negative impact on PageRank.
- Accidental exclusions deprive your site of indexable pages and block the transmission of internal equity.
- You should regularly audit the Search Console to detect unintentional blocks (noindex, robots.txt, mispositioned canonicals).
- Google does not penalize a site for a high number of excluded pages—only their nature matters.
- A page blocked by robots.txt cannot pass any PageRank, even if it receives external backlinks.
SEO Expert opinion
Is this statement consistent with what we observe in practice?
Yes, broadly speaking. It has been known for years that Google does not penalize a site for having a large number of non-indexed pages. Large e-commerce sites naturally generate thousands of excluded URLs through facets, filters, pagination—and that does not deteriorate their rankings as long as strategic pages remain indexable.
But be cautious: stating that excluded pages 'do not influence PageRank' is only true if they are excluded after discovery. A page blocked upstream by robots.txt is not even crawled—so it cannot pass any equity, even if it receives backlinks. This is a nuance that Google does not detail here.
What uncertainties remain regarding this statement?
Google says nothing about the impact of crawl budget. If Google spends its time discovering and re-crawling thousands of excluded pages, it monopolizes resources that could be used elsewhere. On a large site, this is not negligible. [To be verified]: Google claims to optimize crawling automatically, but our observations show that cleaning up irrelevant URLs often speeds up the indexing of priority content.
Another blind spot: soft 404s and 'detected but not indexed' pages. Google categorizes them as excluded, but their proliferation can signal a quality or structural issue. Saying they do not affect PageRank is technically correct—but they often reveal deeper flaws.
When should you really worry about excluded pages?
When strategic pages are accidentally excluded. If your key product sheets, SEO landing pages, or premium editorial content are marked noindex or blocked by robots.txt, you lose both indexing and the ability to pass equity. That’s where the problem lies.
The other critical case: a site that switches to HTTPS and forgets to properly redirect its old URLs. If Google continues to crawl the HTTP versions and marks them as excluded, that dilutes signals and delays migration. Technically, these pages do not affect PageRank—but they create a background noise that slows everything down.
Practical impact and recommendations
What concrete steps should you take to avoid accidental exclusions?
Audit your Search Console each month. Export the list of excluded pages, sort them by reason for exclusion (noindex, robots.txt, canonical, soft 404), and cross-reference with your crawl plan. Identify the URLs that should not be there.
Then, correct any irrelevant directives. If a strategic page has a noindex, remove it. If robots.txt blocks an entire section you want to index, modify the file. If a canonical points to the wrong URL, adjust it. These technical errors are common after a redesign or migration.
How can you ensure that intentionally excluded pages remain excluded?
Document your exclusion strategy. Explicitly list the types of pages you want to block: facet filters, internal search result pages, tracking URLs, thank-you pages. Then check that the directives (noindex, robots.txt) are correctly in place and coherent.
Do not rely on Google to guess your intentions. If you want a page to stay out of the index, assert it with a clean noindex tag. If you don’t want it to be crawled at all, block it via robots.txt—but be aware it will then pass no internal equity.
What critical mistakes should you absolutely avoid?
Blocking pages through robots.txt while hoping they will transmit internal PageRank. It doesn't work. A page blocked upstream is not crawled, so it cannot distribute anything—even if it receives external backlinks.
Another common mistake: leaving noindex tags lingering after a migration or testing. If you marked an entire section with noindex for an audit, and forget to remove the directive, those pages remain excluded indefinitely. Always check systematically after any technical intervention.
- Monthly export excluded pages from Search Console and identify anomalies
- Remove noindex tags from strategic pages that should not be excluded
- Ensure robots.txt does not block sections you want to index
- Document your exclusion strategy to avoid errors during redesigns
- Never block a page you're trying to exploit the internal PageRank of with robots.txt
- After each migration or technical test, audit noindex and canonical directives
❓ Frequently Asked Questions
Une page bloquée par robots.txt transmet-elle du PageRank interne ?
Un grand nombre de pages exclues pénalise-t-il mon site ?
Faut-il nettoyer toutes les pages marquées comme exclues dans la Search Console ?
Comment savoir si une page est exclue accidentellement ou volontairement ?
Les pages « détectées mais non indexées » affectent-elles le PageRank ?
🎥 From the same video 8
Other SEO insights extracted from this same Google Search Central video · duration 1h08 · published on 24/01/2019
🎥 Watch the full video on YouTube →
💬 Comments (0)
Be the first to comment.