What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 5 questions

Less than a minute. Find out how much you really know about Google search.

🕒 ~1 min 🎯 5 questions

Official statement

To avoid Google confusing two similar content sites, ensure that each site has unique content and does not appear to be a copy of the other.
46:05
🎥 Source video

Extracted from a Google Search Central video

⏱ 57:14 💬 EN 📅 23/01/2018 ✂ 27 statements
Watch on YouTube (46:05) →
Other statements from this video 26
  1. 8:27 L'expérience utilisateur suffit-elle vraiment à contourner Panda ?
  2. 10:11 Faut-il vraiment changer le contenu d'une page à chaque visite pour mieux ranker ?
  3. 11:00 Les redirections 301 transfèrent-elles vraiment tous les signaux SEO vers la nouvelle URL ?
  4. 11:04 Les redirections 301 transfèrent-elles vraiment tous les signaux SEO vers la nouvelle URL ?
  5. 11:38 Les liens internes positionnés en bas de page perdent-ils leur valeur SEO ?
  6. 13:41 Pourquoi le Knowledge Graph disparaît-il après une restructuration de site ?
  7. 16:19 JavaScript, mobile et données structurées : pourquoi Google pousse-t-il ces trois chantiers simultanément ?
  8. 16:21 Pourquoi le rendu JavaScript peut-il torpiller votre visibilité dans Google ?
  9. 19:05 Votre site mobile est-il vraiment équivalent à votre version desktop ?
  10. 19:33 Faut-il vraiment rediriger les produits en rupture définitive vers des alternatives ?
  11. 23:31 Pourquoi les balises canonical sont-elles critiques pour vos sites multilingues ?
  12. 23:53 Comment gérer la canonicalisation des sites multilingues sans perdre votre trafic international ?
  13. 25:40 Comment Google gère-t-il vraiment le contenu dupliqué sur votre site ?
  14. 28:36 Comment signaler efficacement du contenu dupliqué à Google ?
  15. 29:29 Le contenu dupliqué interne est-il vraiment un problème pour votre référencement ?
  16. 32:43 Faut-il vraiment conserver les URLs de produits définitivement retirés du catalogue ?
  17. 33:30 Le défilement infini tue-t-il vraiment votre référencement ?
  18. 34:52 Faut-il supprimer les pages produits en rupture de stock ou les conserver indexées ?
  19. 37:36 La position des liens internes sur la page affecte-t-elle vraiment le classement Google ?
  20. 46:30 Google réécrit-il vraiment vos méta-descriptions comme bon lui semble ?
  21. 47:04 La Search Console cache-t-elle une partie de vos données de trafic ?
  22. 49:34 Les liens dans les PDF transmettent-ils du PageRank et améliorent-ils le classement ?
  23. 54:47 Google utilise-t-il vraiment des scores de lisibilité pour classer vos contenus ?
  24. 55:23 La vitesse de page mobile suffit-elle vraiment à faire décoller votre classement ?
  25. 55:29 La vitesse mobile est-elle vraiment un facteur de classement prioritaire sur Google ?
  26. 179:16 Les données structurées influencent-elles vraiment le classement Google ?
📅
Official statement from (8 years ago)
TL;DR

Google treats sites with similar content as potential duplicates, diluting their respective visibility. To distinguish two legitimately close sites, each domain must provide truly unique content and a distinct editorial identity. This isn't just cosmetic: without clear differentiation, Google may merge signals or ignore one of the two sites.

What you need to understand

Why does Google merge signals from similar sites?

Google does not simply detect exact text duplicates. Its algorithm also evaluates semantic proximity, structure, targeted keywords, and overall architecture. When two sites resemble each other too closely, the engine treats them as variants of a single entity.

This merging can lead to position cannibalization: Google arbitrarily chooses which one to serve in the results, ignoring the other. Backlinks and authority become dispersed instead of concentrated, diluting the organic performance of both domains.

What does Google mean by 'unique content'?

The term 'unique' does not simply mean rephrasing the same ideas. Google expects a distinct editorial angle, different examples, and a clearly differentiated target audience. Two sites on the same topic can coexist without issue if they serve complementary search intents.

For instance, a B2C site and a B2B site in the same industry remain distinct if their vocabulary, tone, and use cases differ. However, two identical sites with just a logo or domain change trigger Google's anti-spam filters.

What signals does Google use to detect similarity?

Google cross-references several technical and editorial fingerprints: similarities in title and meta tags, overlap of keywords in H1-H6, lexical density of content, navigation structure, and the same profile of incoming links. If these elements converge, the engine considers it intentional duplication or an attempt to manipulate.

Owners of multi-product agencies or franchises often fall into this trap by deploying mirror sites with copied content. The result is predictable: neither site performs correctly as Google pits them against each other.

  • Google detects duplicates through structure and semantics, not just through exact text.
  • Similar sites cannibalize their own positions and disperse their authority.
  • Unique content requires a distinct editorial angle, not just a rephrasing.
  • Technical signals (tags, architecture, links) are compared to identify variants of the same site.
  • Franchises and multi-product agencies are particularly vulnerable to this risk.

SEO Expert opinion

Is this recommendation consistent with field observations?

Absolutely. Cases of inter-domain cannibalization are common in agencies: a client launches a new site without closing the old one, or a brand creates sub-brands with nearly identical content. The outcome is systematic: both sites stagnate in the SERPs instead of advancing.

Google applies the same logic here as for internal cannibalization: faced with two pages or sites that are too similar, it arbitrarily chooses, ignores the other, or alternates their visibility. Stability is impossible under these conditions, and SEO performance becomes unpredictable.

What nuances should be added to this directive?

Mueller does not provide a quantitative threshold for defining 'too similar.' Is it 70% identical content? 50%? Google never states this. [To be verified] by progressively testing limits on low-stakes sites before generalizing.

Moreover, some sectors naturally require close content (comparison sites, aggregators, very specialized niche sites). In these cases, differentiation must come through UX, functionality, filters, customer testimonials, and not just through text. Google can distinguish between real added value and spin content.

In what cases does this rule not apply strictly?

Multi-country or multilingual sites escape this logic if each version provides translated and locally adapted content. Google does not consider a .fr version and a .de version as duplicates, provided that hreflang is implemented correctly and the content respects cultural specifics.

Similarly, editorial sites under different licenses (syndication, content partnerships) can coexist if the legal agreements and canonical attributions are clear. But watch out: Google always prioritizes the original source if it is identifiable.

Caution: If you manage multiple sites in the same sector, immediately audit their semantic and technical proximity. The risk of invisible cannibalization is real and may explain inexplicably low SEO performance.

Practical impact and recommendations

What concrete steps should be taken to differentiate two close sites?

Start with a similarity audit: compare the titles, H1s, and the first 200 words of each strategic page. If more than 60% of the vocabulary is identical, you are in the red zone. Then, reframe the editorial angle of each site targeting distinct search intents.

For example, if you have an e-commerce site and a blog site on the same product, ensure that the blog discusses uses, experiences, practical guides while the e-commerce site remains focused on optimized product pages for conversion. This separation of intent avoids any algorithmic confusion.

What mistakes should be absolutely avoided?

Never duplicate title and meta description tags between two sites, even if the topics are close. Google interprets this duplication as an attempt at spamming or as an illegitimate mirror site. Also, vary the internal link anchors and introductory texts.

Avoid pointing identical backlinks to both sites from the same referring domains. This reinforces the perception of duplicates. If you must mention both sites in the same external article, clearly differentiate their positioning and respective added value.

How can I verify that my sites are genuinely distinct in Google's eyes?

Use semantic similarity tools (TF-IDF, LSI analysis) to measure lexical overlap. Also, compare link profiles using Ahrefs or Majestic: if both sites share more than 40% of the same referring domains, Google may treat them as variants.

Test for cannibalization by searching for specific long-tail queries: if both sites appear alternately on the same query without clear logic, it is a sign of algorithmic confusion. In this case, strengthen editorial and technical differentiation before Google penalizes one of the two.

  • Audit the textual similarity of titles, H1s, and the first 200 words of each strategic page.
  • Differentiating editorial angles by targeting complementary search intents, not competing ones.
  • Never duplicate title and meta description tags between the two sites.
  • Avoid pointing identical backlinks to both domains from the same sources.
  • Measure lexical overlap with TF-IDF tools and compare link profiles.
  • Test for cannibalization by searching for long-tail queries and observe which site Google serves.
Differentiating two similar sites requires strategic work on editorial angles, technical structure, and link profiles. These cross-optimizations can be complex to orchestrate alone, especially if business stakes are high. Consulting with a specialized SEO agency allows for an accurate diagnosis, a tailored roadmap, and on-the-ground support to avoid any invisible cannibalization.

❓ Frequently Asked Questions

Deux sites sur la même thématique peuvent-ils coexister sans problème ?
Oui, à condition qu'ils ciblent des intentions de recherche différentes et proposent un angle éditorial distinct. Google ne pénalise pas la proximité thématique, mais la duplication éditoriale et technique.
Quel pourcentage de similarité déclenche un filtre duplicate ?
Google ne communique aucun seuil précis. En pratique, au-delà de 60-70% de recoupement lexical sur les éléments structurants (titles, H1, introduction), le risque de cannibalisation devient élevé.
Les sites multilingues sont-ils concernés par cette règle ?
Non, si le hreflang est correctement implémenté et que chaque version propose un contenu adapté localement. Google distingue les versions linguistiques légitimes des duplicates.
Comment détecter une cannibalisation entre deux de mes sites ?
Recherchez des requêtes longue traîne spécifiques : si les deux sites apparaissent alternativement sans logique claire, c'est un signe de confusion algorithmique. Comparez aussi les profils de liens et la similarité sémantique.
Faut-il fermer l'un des deux sites si la similarité est trop forte ?
Pas nécessairement. Si les deux sites ont une légitimité business, renforcez leur différenciation éditoriale et technique. Si l'un n'a plus de raison d'être, une redirection 301 vers l'autre consolidera l'autorité.
🏷 Related Topics
Content AI & SEO

🎥 From the same video 26

Other SEO insights extracted from this same Google Search Central video · duration 57 min · published on 23/01/2018

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.