What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 5 questions

Less than a minute. Find out how much you really know about Google search.

🕒 ~1 min 🎯 5 questions

Official statement

Noindex-marked pages that contain relevant links can still contribute to content discovery. Googlebot will continue to follow links if they are regularly found through internal links.
10:45
🎥 Source video

Extracted from a Google Search Central video

⏱ 56:44 💬 EN 📅 13/06/2019 ✂ 11 statements
Watch on YouTube (10:45) →
Other statements from this video 10
  1. 1:36 Faut-il vraiment rediriger chaque URL individuellement lors d'un déménagement de site ?
  2. 2:39 Pourquoi l'outil de changement d'adresse bloque-t-il les URL avec paramètres ?
  3. 5:21 Faut-il indexer toutes les variations de produit ou canoniser vers la page principale ?
  4. 14:29 Le contenu masqué dans les menus mobiles est-il vraiment pris en compte pour le SEO ?
  5. 21:31 Les contenus uniques offrent-ils vraiment un avantage SEO mesurable ?
  6. 28:45 Faut-il vraiment recycler la même URL pour vos contenus saisonniers annuels ?
  7. 31:06 Faut-il dupliquer vos images pour chaque version linguistique de votre site ?
  8. 48:52 Google utilise-t-il vraiment des critères de classement différents entre mobile et desktop ?
  9. 74:00 Hreflang sans contenu différencié : pourquoi Google ne garantit-il pas l'affichage distinct des versions ?
  10. 78:40 Faut-il vraiment varier les orthographes d'un mot-clé pour éviter la pénalité bourrage ?
📅
Official statement from (6 years ago)
TL;DR

Google confirms that noindex pages continue to be crawled if they are regularly discovered through internal linking, and their links remain followed. In practical terms, your noindex pagination pages can still contribute to deep content discovery. This nuance challenges the common belief that a noindex tag completely cuts off value transmission — but be cautious, as it does not mean it passes PageRank in the same way as an indexed page.

What you need to understand

Why does Googlebot still crawl pages marked as noindex?

The noindex tag tells Google not to include a page in its index, but it doesn't instruct it to stop visiting the page. If a page regularly receives internal links, Googlebot considers it a part of the site's active architecture and continues to crawl it.

This distinction is crucial for e-commerce sites or blogs with pagination. Category pages 2, 3, 4 often contain products or articles that Google wouldn't discover through page 1 alone — especially if the number of displayed items is limited.

What does this change for content discovery?

A classic scenario: you have 200 products in a category, displayed in pages of 20. Without crawling pagination pages, Google will only see the first 20 products if you block access to subsequent pages or if they are never crawled.

With the noindex maintained and an active internal linking (links like “next page,” “previous,” or page numbers), Googlebot follows these links, discovers the remaining 180 products, and can index them individually if their URLs are crawlable and indexable.

Does this statement apply to all types of noindex?

Mueller specifically talks about pagination lists, not all noindex pages. Behavior can vary depending on whether the noindex page is orphaned, rarely linked, or instead present within the permanent linking structure of the site.

Google adjusts the crawl frequency based on the internal “popularity” of a URL. A pagination page regularly linked from indexed pages will be crawled more often than an isolated noindex page, infrequently updated and poorly linked.

  • Noindex does not prevent crawling if the page is regularly found through active internal links
  • Links present on these pages remain followed, allowing for deep content discovery
  • Behavior varies based on the integration level of the page within the site's internal linking structure
  • This mechanism is particularly relevant for e-commerce pagination and large blog archives

SEO Expert opinion

Is this statement consistent with field observations?

Yes, and it confirms what many of us have observed for years: noindex ≠ nofollow. Server logs show that Googlebot continues to visit well-linked noindex pages, sometimes at a frequency almost identical to that of equivalent indexed pages.

But — and this is where Mueller remains deliberately vague — nothing indicates that these pages pass PageRank in the same way. Google has always been evasive on this point. Tests show contradictory results depending on websites, sectors, and linking structure. [To be verified]

What nuances should be applied to this logic?

The phrasing “regularly found through internal links” is crucial. If your pagination pages are only accessible via JavaScript, or if they receive links only from other noindex pages, the allocated crawl budget will be drastically reduced.

Another point: Mueller does not mention value transmission. A noindex page can serve as a relay to discover content, but there's no guarantee it boosts the ranking of target pages like an indexed page with accumulated PageRank would. Real-world tests show diminished or even zero transmission in some cases. [To be verified]

What cases make this logic counterproductive?

If you multiply noindex pages within your architecture — combinatorial filters, sorting by price/date, minor category variations — you force Google to massively crawl URLs without indexable value. The crawl budget gets diluted, and strategic pages are visited less frequently.

Let’s be honest: many e-commerce sites have thousands of noindex pagination or filter pages that contribute nothing, as the products are already discoverable via shorter paths. In these cases, it’s better to completely block via robots.txt or disable links to save crawl budget.

Warning: If your site generates thousands of combinations of filters or sorting pages, all in noindex but all crawlable, you risk saturating your crawl budget without measurable gain. Audit the actual crawl frequency before generalizing this approach.

Practical impact and recommendations

What should you do concretely on a site with pagination?

First, ask yourself: do my pagination pages actually serve to discover content? If your categories display 10 products per page and you have 300, yes, absolutely. If you show 100 products per page and page 2 hardly ever exists, the debate is closed.

Then, check your server logs to see if Googlebot is actually crawling these pages. If they are noindex but never visited, it means they are poorly linked or Google has decided they aren't worth visiting. Adjust the linking or remove them from the equation.

What mistakes should be avoided in managing noindex on pagination?

A classic mistake: putting pagination pages in noindex but also in nofollow or blocking them via robots.txt. You completely cut off the discovery of deep content. Google no longer follows the links, and your products or articles on page 3+ will never be seen.

Another trap: using noindex on pagination pages that are the only access route to certain content. If a product only appears on page 12 of a category and has no direct link from the menu or sitemap, it will never be indexed if that page 12 is blocked or not crawled.

How can I check if this strategy works on my site?

Install a log analyzer (Botify, OnCrawl, Screaming Frog Log Analyzer) and filter the URLs containing your pagination parameters (“?page=”, “/page/2/”, etc.). Look at the crawl frequency, HTTP status, and discovery depth.

Then compare it with the indexing coverage of the products or articles listed on those pages. If you find that deep content is never indexed despite regular crawling of pagination pages, the problem lies elsewhere — perhaps a lack of internal popularity, a saturated crawl budget, or content deemed too weak.

  • Maintain pagination pages in noindex, index-follow (or without explicit nofollow)
  • Ensure that the internal linking to these pages is active and regular (links like “next page,” visible page numbers)
  • Check in the server logs that Googlebot is effectively crawling these URLs with a reasonable frequency
  • Audit the indexing coverage of listed deep content to confirm their discovery
  • Avoid unnecessarily multiplying crawlable noindex pages if they do not contribute to discovery (redundant combinatorial filters, sorting without added value)
  • Consider a comprehensive XML sitemap for strategic deep content, in addition to pagination linking
The noindex tag on pagination can coexist with active crawling and effective deep content discovery, provided that the internal linking is robust and the crawl budget is correctly allocated. This is not a one-size-fits-all solution — it requires a precise audit of the architecture, logs, and actual indexing. If your site generates tens of thousands of filter or sorting pages, this approach can quickly become counterproductive. In these complex cases, personalized support from a specialized SEO agency can help calibrate the right strategy between noindex, robots.txt, canonicals, and optimized linking, depending on your actual crawl budget and business priorities.

❓ Frequently Asked Questions

Est-ce que les pages noindex transmettent du PageRank ?
Google n'a jamais confirmé explicitement que les pages noindex transmettent du PageRank de la même manière que les pages indexées. Les observations terrain montrent des résultats variables, ce qui suggère une transmission atténuée ou conditionnelle selon l'architecture du site.
Faut-il mettre les pages de pagination en noindex ou les laisser indexées ?
Cela dépend de votre architecture. Si elles génèrent du contenu dupliqué ou diluent le crawl budget sans apporter de valeur SEO propre, le noindex est pertinent. Si elles sont peu nombreuses et bien optimisées, vous pouvez les laisser indexées avec des canoniques ou des balises rel=next/prev.
Googlebot crawle-t-il toutes les pages noindex de mon site ?
Non, seulement celles qui sont régulièrement trouvées via le maillage interne et jugées utiles pour la découverte de contenu. Une page noindex orpheline ou rarement liée sera rapidement déprioritisée dans le crawl budget.
Peut-on bloquer une page de pagination via robots.txt tout en gardant un noindex ?
Non, c'est contradictoire et contre-productif. Si vous bloquez via robots.txt, Google ne peut pas accéder à la page pour lire la balise noindex, et il ne suivra pas les liens qu'elle contient. Choisissez l'un ou l'autre, pas les deux.
Comment savoir si mes pages noindex contribuent réellement à la découverte de contenu ?
Analysez vos logs serveur pour vérifier la fréquence de crawl de ces pages, puis comparez avec l'indexation des contenus qu'elles lient. Si Google crawle régulièrement ces pages mais que les contenus profonds ne sont pas indexés, le problème se situe ailleurs (qualité, popularité, crawl budget saturé).
🏷 Related Topics
Domain Age & History Content Crawl & Indexing AI & SEO Links & Backlinks Pagination & Structure

🎥 From the same video 10

Other SEO insights extracted from this same Google Search Central video · duration 56 min · published on 13/06/2019

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.