What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 5 questions

Less than a minute. Find out how much you really know about Google search.

🕒 ~1 min 🎯 5 questions

Official statement

Pages marked as excluded in Google Search Console do not influence PageRank. These pages are often ignored unless they are accidentally excluded through 'noindex' tags or directives in robots.txt. In that case, it is necessary to remove those directives for the pages to be indexed.
1:52
🎥 Source video

Extracted from a Google Search Central video

⏱ 1h08 💬 EN 📅 24/01/2019 ✂ 9 statements
Watch on YouTube (1:52) →
Other statements from this video 8
  1. 5:31 Un HTML correct améliore-t-il vraiment votre classement SEO ?
  2. 9:17 Les canonicals suffisent-ils vraiment à gérer les doublons sans pénalité SEO ?
  3. 25:47 La balise noindex bloque-t-elle vraiment l'indexation de vos pages stratégiques ?
  4. 31:36 Les signaux sociaux influencent-ils vraiment le classement dans Google ?
  5. 34:19 Le PageRank influence-t-il encore vraiment le classement Google en SEO ?
  6. 39:58 L'achat de liens et les échanges de backlinks conduisent-ils vraiment à des pénalités ?
  7. 55:24 Les pages AMP exclues de l'index signalent-elles vraiment une mauvaise implémentation ?
  8. 67:02 Le contenu de qualité suffit-il vraiment à bien se positionner dans Google ?
📅
Official statement from (7 years ago)
TL;DR

Google states that pages marked as excluded in the Search Console have no impact on your site's PageRank. Only accidental exclusions—via noindex or robots.txt—are problematic and block indexing. For SEO, this means it's crucial to distinguish between intentional exclusions and technical errors that deprive strategic pages of their ability to pass link equity.

What you need to understand

What does 'excluded' really mean in the Search Console?

The Search Console groups under 'excluded' all the pages that Google has discovered but does not index. This catch-all category mixes very different situations: pages blocked by robots.txt, marked noindex, detected as duplicates, or simply ignored because deemed worthless.

The problem is that this vague terminology creates confusion. An 'excluded' page can be a page that you want to exclude—such as a thank-you page or a facet filter—or a page you wanted to index but that Google has disregarded. The distinction is not trivial.

Why does Google specify that these pages do not affect PageRank?

Google is responding here to a widespread belief: that thousands of pages excluded in the Search Console would dilute internal PageRank or penalize the site. This fear is unfounded if the exclusions are intentional and legitimate.

On the other hand, if strategic pages are accidentally excluded—due to a forgotten noindex or a poorly calibrated robots.txt block—they cannot be indexed or pass their equity. Here, the SEO impact becomes real, even if it's not the number of excluded pages that is the issue, but their nature.

How can you distinguish between an intentional exclusion and a technical error?

An intentional exclusion results from a directive that you put in place: noindex tag, canonical pointing to another URL, targeted robots.txt block. An accidental exclusion is when a page you want to index ends up on this list without your intent.

In practical terms, a thorough audit of the Search Console is necessary. Cross-reference the list of excluded URLs with your crawl plan and editorial priorities. If high-potential pages—product sheets, landing pages—appear as excluded, it’s a warning signal.

  • Intentionally excluded pages (filters, technical pages) have no negative impact on PageRank.
  • Accidental exclusions deprive your site of indexable pages and block the transmission of internal equity.
  • You should regularly audit the Search Console to detect unintentional blocks (noindex, robots.txt, mispositioned canonicals).
  • Google does not penalize a site for a high number of excluded pages—only their nature matters.
  • A page blocked by robots.txt cannot pass any PageRank, even if it receives external backlinks.

SEO Expert opinion

Is this statement consistent with what we observe in practice?

Yes, broadly speaking. It has been known for years that Google does not penalize a site for having a large number of non-indexed pages. Large e-commerce sites naturally generate thousands of excluded URLs through facets, filters, pagination—and that does not deteriorate their rankings as long as strategic pages remain indexable.

But be cautious: stating that excluded pages 'do not influence PageRank' is only true if they are excluded after discovery. A page blocked upstream by robots.txt is not even crawled—so it cannot pass any equity, even if it receives backlinks. This is a nuance that Google does not detail here.

What uncertainties remain regarding this statement?

Google says nothing about the impact of crawl budget. If Google spends its time discovering and re-crawling thousands of excluded pages, it monopolizes resources that could be used elsewhere. On a large site, this is not negligible. [To be verified]: Google claims to optimize crawling automatically, but our observations show that cleaning up irrelevant URLs often speeds up the indexing of priority content.

Another blind spot: soft 404s and 'detected but not indexed' pages. Google categorizes them as excluded, but their proliferation can signal a quality or structural issue. Saying they do not affect PageRank is technically correct—but they often reveal deeper flaws.

When should you really worry about excluded pages?

When strategic pages are accidentally excluded. If your key product sheets, SEO landing pages, or premium editorial content are marked noindex or blocked by robots.txt, you lose both indexing and the ability to pass equity. That’s where the problem lies.

The other critical case: a site that switches to HTTPS and forgets to properly redirect its old URLs. If Google continues to crawl the HTTP versions and marks them as excluded, that dilutes signals and delays migration. Technically, these pages do not affect PageRank—but they create a background noise that slows everything down.

Warning: A large number of 'detected but not indexed' pages may indicate that Google considers your content to be of low quality or too redundant. This is not a PageRank issue, it is a problem of overall perception of your site.

Practical impact and recommendations

What concrete steps should you take to avoid accidental exclusions?

Audit your Search Console each month. Export the list of excluded pages, sort them by reason for exclusion (noindex, robots.txt, canonical, soft 404), and cross-reference with your crawl plan. Identify the URLs that should not be there.

Then, correct any irrelevant directives. If a strategic page has a noindex, remove it. If robots.txt blocks an entire section you want to index, modify the file. If a canonical points to the wrong URL, adjust it. These technical errors are common after a redesign or migration.

How can you ensure that intentionally excluded pages remain excluded?

Document your exclusion strategy. Explicitly list the types of pages you want to block: facet filters, internal search result pages, tracking URLs, thank-you pages. Then check that the directives (noindex, robots.txt) are correctly in place and coherent.

Do not rely on Google to guess your intentions. If you want a page to stay out of the index, assert it with a clean noindex tag. If you don’t want it to be crawled at all, block it via robots.txt—but be aware it will then pass no internal equity.

What critical mistakes should you absolutely avoid?

Blocking pages through robots.txt while hoping they will transmit internal PageRank. It doesn't work. A page blocked upstream is not crawled, so it cannot distribute anything—even if it receives external backlinks.

Another common mistake: leaving noindex tags lingering after a migration or testing. If you marked an entire section with noindex for an audit, and forget to remove the directive, those pages remain excluded indefinitely. Always check systematically after any technical intervention.

  • Monthly export excluded pages from Search Console and identify anomalies
  • Remove noindex tags from strategic pages that should not be excluded
  • Ensure robots.txt does not block sections you want to index
  • Document your exclusion strategy to avoid errors during redesigns
  • Never block a page you're trying to exploit the internal PageRank of with robots.txt
  • After each migration or technical test, audit noindex and canonical directives
Excluded pages do not affect your PageRank—except when the wrong pages are excluded. The key is to master your indexing directives and regularly cross-reference the Search Console with your editorial strategy. If managing these optimizations seems complex to handle alone, or if your site generates thousands of technical URLs that are difficult to prioritize, working with a specialized SEO agency can help you structure a rigorous crawl plan and avoid costly errors during migrations.

❓ Frequently Asked Questions

Une page bloquée par robots.txt transmet-elle du PageRank interne ?
Non. Une page bloquée par robots.txt n'est pas crawlée par Google, donc elle ne peut transmettre aucun jus, même si elle reçoit des backlinks externes.
Un grand nombre de pages exclues pénalise-t-il mon site ?
Non, Google ne pénalise pas un site pour un nombre élevé de pages exclues. Seul leur type compte : si des pages stratégiques sont exclues par erreur, tu perds de l'indexation et du jus.
Faut-il nettoyer toutes les pages marquées comme exclues dans la Search Console ?
Non. Si ce sont des exclusions volontaires (filtres, paginations, pages techniques), il n'y a rien à faire. Seules les exclusions accidentelles nécessitent une correction.
Comment savoir si une page est exclue accidentellement ou volontairement ?
Croise la liste des URLs exclues avec ton plan de crawl et tes priorités éditoriales. Si une page stratégique apparaît comme exclue, vérifie si elle porte un noindex ou un blocage robots.txt non intentionnel.
Les pages « détectées mais non indexées » affectent-elles le PageRank ?
Non, mais leur multiplication peut signaler que Google juge ton contenu de faible qualité ou trop redondant. Ce n'est pas un problème de PageRank, mais un signal d'alerte sur la perception globale de ton site.
🏷 Related Topics
Domain Age & History Crawl & Indexing Links & Backlinks Search Console

🎥 From the same video 8

Other SEO insights extracted from this same Google Search Central video · duration 1h08 · published on 24/01/2019

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.