What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 3 questions

Less than 30 seconds. Find out how much you really know about Google search.

🕒 ~30s 🎯 3 questions 📚 SEO Google

Official statement

User comments visible in the rendered HTML do not need to be hidden from Google. Google is capable of identifying user-generated content, and it neither benefits nor harms SEO. Trying to hide them may create more serious issues (accidentally blocking critical resources).
10:35
🎥 Source video

Extracted from a Google Search Central video

⏱ 19:34 💬 EN 📅 11/06/2020 ✂ 5 statements
Watch on YouTube (10:35) →
Other statements from this video 4
  1. 7:08 Faut-il vraiment limiter le nombre de ressources HTTP par page pour le SEO ?
  2. 13:49 Un taux de crawl faible est-il vraiment un problème pour votre SEO ?
  3. 14:51 Comment débloquer une page blanche dans Google avec la méthode de bissection ?
  4. 18:01 Un en-tête noindex sur une API empêche-t-il vraiment Googlebot de rendre la page ?
📅
Official statement from (5 years ago)
TL;DR

Google claims to natively identify user-generated content (comments, forums, reviews) without needing to hide it from crawling. Attempting to hide it via robots.txt, aggressive lazy loading, or JavaScript can create technical issues (blocking critical resources, indexing errors). For an SEO practitioner, this means stop complicating things with convoluted technical solutions and accept that Google can differentiate between your editorial content and that of your users.

What you need to understand

Does Google really distinguish between user content and editorial content?

Martin Splitt claims that Google natively identifies user comments and treats them differently from the main content. In practice, the algorithm can detect sections of comments, forums, customer reviews without you having to specifically mark them up or exclude them from crawling.

This distinguishing ability likely relies on several signals: repetitive HTML structure, presentation patterns (avatars, dates, reply buttons), density of short and conversational content, implicit metadata. Google has never detailed exactly how it does this — and that's where it gets interesting for us.

Why is this statement being made now?

Because too many sites implement complex technical solutions to hide comments from Google, often based on persistent SEO myths: fear of content dilution, concern over spam, desire to control internal PageRank. These practices cause more problems than they solve.

The main risk? Accidentally blocking critical resources. A misconfigured robots.txt that aims to exclude /comments/ could also block /comments.css or /comments.js if those files are used elsewhere. An overly aggressive lazy loading may prevent Google from crawling valid URLs found in the comments.

What impact on ranking if I keep everything visible?

Splitt states that leaving comments visible in the rendered HTML neither benefits nor harms SEO. In other words, their presence is neutral — Google sees them, processes them, but doesn't use them to evaluate the quality of your main page.

This neutrality aligns with our observations: a blog post with 200 comments does not rank better than a similar post without comments. However, a post with comments rich in natural keywords may attract long-tail traffic if Google indexes those segments — and that’s where it gets a bit complicated.

  • Google automatically identifies user comment sections without specific markup
  • Hiding comments creates more risks than benefits (blocking resources, crawl errors)
  • Their presence is neutral for the main page's ranking
  • No need for robots.txt, noindex, or lazy loading to exclude them
  • Focus your efforts on the quality of your editorial content, not on marginal technical optimizations

SEO Expert opinion

Is this claim consistent with real-world observations?

Yes and no. On mature platforms (WordPress, phpBB forums, standard e-commerce sites), Google seems indeed capable of distinguishing comments from main content. We see this in featured snippets: Google rarely extracts text from comments to answer a query, preferring the body of the article.

However, on custom sites with non-standard HTML structures, or SPAs (Single Page Applications) where everything is rendered in JavaScript, this distinction becomes less reliable. I've seen cases where Google massively indexed comment threads as standalone pages, creating unintentional duplicate content. [To verify]: Google has never published benchmarks showing the accuracy rate of this detection across different site types.

In what cases does this rule not apply?

The first obvious case: mass unmoderated spam. If your comments are saturated with links to dubious sites, Google may indeed degrade the overall perception of your page. Splitt says it’s neutral, but in reality, a site that allows 500 spam comments per day will take algorithmic hits.

The second case: sites where comments are the main content (Reddit, Stack Overflow, Quora). Here, Google does not apply this separation logic — everything is treated as editorial content. Splitt's statement clearly targets blogs, news sites, e-commerce, not conversational platforms.

What is the real nuance that Google does not mention?

Google claims that keeping comments visible is neutral for ranking, but fails to mention that some user signals may still come into play: time spent on the page extended by reading comments, reduced bounce rate if comments add value, boosted organic CTR if rich snippets display a number of comments.

In other words, comments do not directly impact the content algorithm, but may indirectly influence through behavioral signals. Google will never publicly make this distinction because it opens the door for manipulation. [To verify]: no official data on how these signals weigh in the overall algorithm.

Attention: If you decide to keep everything visible, ensure that your anti-spam moderation is robust. An influx of low-quality comments may degrade user experience, impacting your Core Web Vitals (CLS if comments load late) and your behavioral metrics.

Practical impact and recommendations

What should I practically do with existing comments?

If you currently have robots.txt rules blocking /comments/ or noindex directives on these sections, consider removing them. But do it gradually: remove the block on a sample of pages, monitor Search Console for 2-3 weeks to detect any anomalies (crawl spikes, 404 errors, unwanted massive indexing).

For new sites or redesigns, the recommendation is simple: do not create a specific exclusion mechanism for comments. Leave them in the rendered HTML, and Google will handle it. Focus your energy on qualitative moderation rather than technical gymnastics.

What mistakes should you absolutely avoid?

The classic mistake: using ultra-aggressive lazy-loading JavaScript that only loads comments on scroll, hoping Google won’t see them. Not only does Google now execute JS (so it will see them anyway), but you also risk blocking critical resources if your comment script shares dependencies with other modules.

Another trap: attempting to mark up comments with schema.org Comment to “help” Google. This is unnecessary in 99% of cases — Google already detects the structure. Worse, poorly implemented markup can generate errors in Search Console and clutter your rich snippets. Keep it simple.

How can I check if my implementation is correct?

Use the URL inspection tool in Search Console to see the rendered HTML as Google sees it. If your comments appear in the rendered DOM, you're good. If you see empty blocks or JS errors, you have a problem — but it's probably not specifically related to the comments.

Also, check that your comment URLs (e.g., /article#comment-123) are not generating massive indexing in Google. If they are, use the canonical tag to point to the main article URL, not a brutal noindex that risks creating inconsistencies.

  • Gradually remove robots.txt rules blocking /comments/ or /reviews/
  • Remove noindex directives on user sections if they exist
  • Avoid aggressive lazy loading that blocks comments from initial rendering
  • Do not over-optimize with schema.org Comment unless you have a specific use case (rich snippets reviews)
  • Check the rendered HTML in Search Console to confirm that Google sees the comments
  • Monitor crawl and indexing metrics after modifications to detect anomalies
In summary: leave your comments alone. Google knows how to handle them, and your efforts are better invested elsewhere — editorial quality, internal linking, technical optimization of critical elements. If your site manages millions of comments or has atypical HTML structures, these optimizations may become complex to orchestrate alone. In this case, engaging an SEO agency specialized in auditing your implementation and supporting you with a user content strategy may be a wise investment.

❓ Frequently Asked Questions

Les commentaires utilisateurs comptent-ils dans le contenu total de ma page pour le ranking ?
Non, selon Google ils sont identifiés séparément et traités comme du contenu utilisateur, donc neutres pour l'évaluation qualitative de votre contenu éditorial principal. Leur présence n'améliore ni ne détériore votre positionnement direct.
Dois-je utiliser du schema.org Comment pour aider Google à distinguer les commentaires ?
Non, ce n'est pas nécessaire. Google détecte automatiquement les commentaires via la structure HTML et les patterns de présentation. Le balisage schema.org peut être utile pour les rich snippets d'avis produits, mais pas pour des commentaires de blog standards.
Si je bloque les commentaires avec robots.txt, est-ce que je risque de bloquer autre chose par accident ?
Oui, c'est le risque principal soulevé par Splitt. Un robots.txt qui vise /comments/ peut bloquer des CSS ou JS partagés, créant des erreurs de rendu et impactant l'indexation de ressources critiques. Mieux vaut ne rien bloquer du tout.
Les commentaires spammés peuvent-ils pénaliser mon site même si Google les identifie comme contenu utilisateur ?
Oui, indirectement. Un site saturé de spam dégrade l'expérience utilisateur, ce qui impacte les métriques comportementales (taux de rebond, temps sur page) et peut nuire à votre réputation globale. La modération reste essentielle.
Faut-il utiliser lazy-loading pour les commentaires afin de ne pas ralentir le chargement de la page ?
Le lazy-loading pour des raisons de performance est acceptable, mais évitez de le faire pour cacher les commentaires de Google. Assurez-vous que le lazy-loading ne bloque pas le rendu initial et que Google peut quand même accéder au contenu via le HTML rendu.
🏷 Related Topics
Domain Age & History Content AI & SEO Web Performance

🎥 From the same video 4

Other SEO insights extracted from this same Google Search Central video · duration 19 min · published on 11/06/2020

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.