What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 3 questions

Less than 30 seconds. Find out how much you really know about Google search.

🕒 ~30s 🎯 3 questions 📚 SEO Google

Official statement

To reliably prevent a page from being indexed in Google Search, you must use the meta robots 'noindex' tag rather than robots.txt, as the latter is not a foolproof method against indexation.
🎥 Source video

Extracted from a Google Search Central video

💬 EN 📅 30/06/2022 ✂ 14 statements
Watch on YouTube →
Other statements from this video 13
  1. Robots.txt bloque-t-il vraiment l'indexation de vos pages ?
  2. La balise meta 'none' est-elle vraiment l'équivalent de noindex + nofollow ?
  3. Peut-on bloquer l'indexation de répertoires entiers via des modules serveur plutôt que robots.txt ?
  4. Faut-il vraiment indexer les pages de connexion de votre site ?
  5. Faut-il vraiment préférer rel=canonical à noindex pour les contenus anciens ?
  6. La balise noarchive empêche-t-elle réellement Google d'archiver vos pages ?
  7. Faut-il bloquer les snippets avec nosnippet pour protéger son contenu sensible ?
  8. Faut-il vraiment utiliser max-snippet et max-image-preview pour contrôler l'affichage dans les SERP ?
  9. Faut-il privilégier l'attribut nofollow individuel ou la balise meta robots nofollow pour contrôler le PageRank ?
  10. Pourquoi Google refuse-t-il de créer de nouvelles balises meta robots ?
  11. Comment bloquer l'indexation de PDFs et fichiers non-HTML sans accès aux headers HTTP ?
  12. Pourquoi robots.txt bloque-t-il vraiment les images et vidéos mais pas les pages web ?
  13. Comment Google transforme-t-il vraiment vos PDFs en contenu indexable ?
📅
Official statement from (3 years ago)
TL;DR

Gary Illyes states that robots.txt does not guarantee blocking indexation in Google Search. Only the meta robots 'noindex' tag provides reliable protection. This technical distinction is significant: a robots.txt file prevents crawling, but does not prevent Google from indexing a URL discovered through other means.

What you need to understand

Why doesn't robots.txt reliably block indexation?

The robots.txt file is a crawl directive mechanism, not an indexation mechanism. It tells robots whether or not they can explore a URL. But if Google discovers this URL through an external backlink or a mention somewhere else, it can index it without ever crawling it.

Result: you find pages in the index with just a simple URL and a generic snippet, with no visible content. This is exactly what happens when you block a page with robots.txt that receives links.

What is the concrete difference between noindex and robots.txt?

The noindex is an indexation instruction. For it to be taken into account, Google must be able to crawl the page and read the meta tag or HTTP header. It's a guarantee: if Googlebot sees the noindex, the page is removed from the index or never enters it.

Conversely, robots.txt prevents access to content but not necessarily to the URL itself. If this URL is referenced elsewhere, Google can decide to index it with the partial information it has.

In which cases does this confusion cause problems?

The classic case: a site blocks sensitive pages (admin, staging, UTM parameters) via robots.txt thinking it will make them invisible. These pages sometimes continue to appear in the SERPs with a truncated snippet.

Another common example: obsolete landing pages or test pages blocked by robots.txt but still linked from old articles. They remain indexed as long as noindex is not applied directly.

  • Robots.txt controls crawl, not indexation
  • Noindex controls indexation permanently
  • A URL blocked by robots.txt can still appear in the index if it receives links
  • To guarantee exclusion from the index, Googlebot must be able to crawl the page and read the noindex
  • Blocking with robots.txt and noindex simultaneously is contradictory and ineffective

SEO Expert opinion

Is this statement consistent with real-world observations?

Yes, completely. It's even a classic of SEO audits: dozens of indexed pages with the message "No information available for this page" because they're blocked in robots.txt but continue to receive backlinks.

The paradox? Some SEOs still use robots.txt to "hide" sensitive content, when that's exactly the opposite of what they should do. This confusion probably comes from a time when tools were less explicit about the distinction between crawl and indexation.

What nuances should be added to this directive?

Gary Illyes doesn't specify one essential point: for noindex to work, you must allow crawling. If you block a page in robots.txt AND add a noindex tag to it, Google will never see the tag. The page may therefore remain indexed indefinitely.

Second nuance — timing. A page with noindex can take several weeks to leave the index if it is rarely crawled or has many backlinks. It's not instant, contrary to what some imagine. [To verify]: Google has never communicated an official timeframe for complete desindexation of a noindex page.

What should you do if you want to block both crawl and indexation?

That's where it gets tricky. If you really want to forbid access AND guarantee the absence of indexation, the proper solution is dual: noindex + server authentication (password, IP whitelisting). Or else noindex until the page leaves the index, then robots.txt blocking.

But in 90% of cases, a clean noindex is sufficient. Robots.txt blocking is rarely necessary for indexation reasons — it mainly serves to save crawl budget on unnecessary sections.

Warning: Never combine robots.txt Disallow and noindex on the same URL. You would prevent Google from seeing your desindexation instruction.

Practical impact and recommendations

What should you concretely do on a production site?

First reflex: audit pages blocked by robots.txt that still appear in the index. Use the site:yourdomain.com command in Google and compare with your robots.txt file. Any URL indexed despite a Disallow is a potential vulnerability.

Next, list all pages you want to genuinely exclude from the index. For each one, apply a noindex via meta robots or X-Robots-Tag HTTP header. Verify that these pages are NOT blocked in robots.txt, otherwise Google will never see the directive.

What errors should you avoid in managing indexation?

Error #1: blocking with robots.txt thinking it will desindex. It won't desindex. It just prevents crawling, and sometimes keeps the URL in the index if it receives links.

Error #2: applying noindex + robots.txt on the same page. Result: Google cannot crawl to read the noindex. The page may remain potentially indexed with an empty snippet.

Error #3: believing that noindex acts instantly. It doesn't. Desindexation depends on crawl frequency and the "strength" of external signals (backlinks, sitemap, etc.).

How do you verify that your configuration is correct?

Use Google Search Console, URL Inspection tool. Enter the URL in question and verify three points: is it crawlable? Is the noindex detected? What is its current indexation status?

If the page is indexed despite recent noindex, wait. If it's indexed and noindex is not detected, check your robots.txt. If noindex is detected but the page still present after several weeks, [To verify] there may be a problem of contradictory signals (sitemap, massive internal links).

  • Remove all Disallow directives from robots.txt for pages you want to desindex
  • Add a <meta name="robots" content="noindex"> tag in the <head> of each page to exclude
  • Verify in Search Console that noindex is properly detected
  • Remove these pages from the XML sitemap to avoid contradictory signals
  • Monitor indexation with site:yourdomain.com every 2-3 weeks
  • Keep crawl authorized until complete desindexation, then block if necessary
Managing indexation relies on a simple but crucial distinction: robots.txt controls crawl, noindex controls indexation. To guarantee that a page is never indexed, you must allow its crawl and apply noindex to it. These mechanisms may seem simple in theory, but their implementation at scale on a complex site requires rigor and expertise. If your technical architecture comprises thousands of pages, dynamic rules, or specific server configurations, it may be wise to call upon a specialized SEO agency to audit your configuration and avoid costly visibility errors.

❓ Frequently Asked Questions

Peut-on utiliser robots.txt ET noindex sur la même page ?
Non, c'est contradictoire. Si vous bloquez le crawl par robots.txt, Google ne pourra jamais voir la balise noindex. La page risque de rester indexée avec un snippet vide.
Combien de temps faut-il pour qu'une page en noindex sorte de l'index ?
Cela dépend de la fréquence de crawl de la page. En général, quelques jours à quelques semaines. Les pages avec beaucoup de backlinks peuvent prendre plus de temps.
Si je bloque une page dans robots.txt, pourquoi apparaît-elle encore dans Google ?
Parce que robots.txt empêche le crawl, pas l'indexation. Si Google découvre l'URL via un lien externe, il peut l'indexer sans jamais la crawler, d'où l'apparition d'un snippet générique.
Faut-il retirer une page en noindex du sitemap XML ?
Oui, absolument. Inclure une page en noindex dans le sitemap envoie des signaux contradictoires à Google. Retirez-la pour accélérer la désindexation.
Le noindex fonctionne-t-il aussi en en-tête HTTP ?
Oui, l'en-tête HTTP X-Robots-Tag: noindex a exactement le même effet que la balise meta. C'est même la seule solution pour les fichiers PDF ou autres ressources non-HTML.
🏷 Related Topics
Domain Age & History Crawl & Indexing AI & SEO

🎥 From the same video 13

Other SEO insights extracted from this same Google Search Central video · published on 30/06/2022

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.