What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 3 questions

Less than 30 seconds. Find out how much you really know about Google search.

🕒 ~30s 🎯 3 questions 📚 SEO Google

Official statement

If a site that has been closed for a year is relaunched on a new domain and its content has been copied by scraping sites, Google will likely not penalize the legitimate site for duplication. The algorithm generally detects who the original author is. If in doubt, report spam sites via the Spam Report.
50:14
🎥 Source video

Extracted from a Google Search Central video

⏱ 1h14 💬 EN 📅 04/06/2020 ✂ 44 statements
Watch on YouTube (50:14) →
Other statements from this video 43
  1. 2:22 Pourquoi votre site a-t-il perdu du trafic après une Core Update sans avoir fait d'erreur ?
  2. 2:22 Les Core Web Vitals vont-ils vraiment bouleverser votre stratégie SEO ?
  3. 3:50 Une baisse de classement après une Core Update signifie-t-elle vraiment un problème avec votre site ?
  4. 3:50 Faut-il vraiment attendre avant d'optimiser les Core Web Vitals ?
  5. 3:50 Pourquoi Google repousse-t-il la migration complète vers le Mobile-First Index ?
  6. 7:07 Google peut-il vraiment repousser le Mobile-First Indexing indéfiniment ?
  7. 11:00 Pourquoi Google ne canonicalise-t-il pas les URLs avec fragments dans les sitelinks et rich results ?
  8. 11:00 Les URLs avec fragments (#) dans Search Console : faut-il revoir votre stratégie de tracking et d'analyse ?
  9. 14:34 Pourquoi les chiffres entre Analytics, Search Console et My Business ne correspondent-ils jamais ?
  10. 14:35 Pourquoi vos métriques Google ne concordent-elles jamais entre Search Console, Analytics et Business Profile ?
  11. 16:37 Comment sont vraiment comptabilisés les clics FAQ dans Search Console ?
  12. 18:44 Les accordéons mobile et desktop sont-ils vraiment neutres pour le SEO ?
  13. 18:44 Le contenu masqué par accordéon mobile est-il vraiment indexé comme du contenu visible ?
  14. 29:45 Le rel=canonical via HTTP header fonctionne-t-il vraiment encore ?
  15. 30:09 L'en-tête HTTP rel=canonical fonctionne-t-il vraiment pour gérer les contenus dupliqués ?
  16. 31:00 Pourquoi Search Console affiche-t-il encore 'PC Googlebot' sur des sites récents alors que le Mobile-First Index est censé être la norme ?
  17. 31:02 Mobile-First Indexing par défaut : pourquoi Search Console affiche-t-il encore desktop Googlebot ?
  18. 33:28 Pourquoi Google insiste-t-il sur le contexte textuel dans les feedbacks Search Console ?
  19. 33:31 Les outils Search Console suffisent-ils vraiment à résoudre vos problèmes d'indexation ?
  20. 33:59 Pourquoi vos pages ne s'indexent-elles toujours pas après 60 jours dans Search Console ?
  21. 37:24 Pourquoi Google indexe-t-il parfois HTTP au lieu de HTTPS malgré la migration SSL ?
  22. 37:53 Faut-il vraiment cumuler redirections 301 ET canonical pour une migration HTTPS ?
  23. 39:16 Pourquoi votre sitemap échoue dans Search Console et comment débloquer réellement la situation ?
  24. 41:29 Votre marque disparaît des SERP sans raison : le feedback Google peut-il vraiment résoudre le problème ?
  25. 44:07 Faut-il privilégier un sous-domaine ou un nouveau domaine pour lancer un service ?
  26. 44:34 Sous-domaine ou nouveau domaine : pourquoi Google refuse-t-il de trancher pour le SEO ?
  27. 44:34 Les pénalités Google se propagent-elles vraiment entre domaine et sous-domaines ?
  28. 45:27 Les pénalités Google se propagent-elles vraiment entre domaine et sous-domaines ?
  29. 48:24 Faut-il vraiment ignorer le PageRank dans le choix entre domaine et sous-domaine ?
  30. 48:33 Les liens entre domaine racine et sous-domaines transmettent-ils réellement du PageRank ?
  31. 49:58 Faut-il vraiment s'inquiéter du contenu dupliqué par scraping ?
  32. 50:14 Faut-il vraiment signaler chaque URL de scraping via le Spam Report pour obtenir une action de Google ?
  33. 57:15 Faut-il vraiment rapporter le spam URL par URL pour aider Google ?
  34. 58:57 Pourquoi Google refuse-t-il d'afficher vos FAQ en rich results malgré un balisage parfait ?
  35. 59:54 Pourquoi Google n'affiche-t-il pas vos FAQ rich results malgré un balisage parfait ?
  36. 65:15 Peut-on ajouter des FAQ sur ses pages uniquement pour gagner des rich results en SEO ?
  37. 65:45 Peut-on ajouter une FAQ uniquement pour obtenir le rich result sans risquer de pénalité ?
  38. 67:27 Faut-il encore optimiser les balises rel=next/prev pour la pagination ?
  39. 67:58 Faut-il vraiment soumettre toutes les pages paginées dans le sitemap XML ?
  40. 70:10 Faut-il vraiment indexer toutes les pages de catégories pour optimiser son crawl budget ?
  41. 70:18 Faut-il vraiment arrêter de mettre les pages catégories en noindex ?
  42. 72:04 Le nombre de fichiers JavaScript ralentit-il vraiment l'indexation Google ?
  43. 72:24 Googlebot rend-il vraiment tout le JavaScript en une seule passe ?
📅
Official statement from (5 years ago)
TL;DR

Google states that relaunching a closed site on a new domain typically does not trigger a duplicate penalty, even if scrapers have copied the content during the downtime. The algorithm is supposed to identify the legitimate original author. In practice, this means that migrating or reactivating archived content should not expose you to a duplicate filter — but if in doubt, reporting spam sites remains the only concrete action suggested.

What you need to understand

What does this statement really mean for a site restarting its activities?

When a site remains closed for an extended period, it's not uncommon for opportunistic scrapers to siphon its content and republish it elsewhere. The classic fear: relaunching the original domain or migrating content to a new domain could trigger a duplicate content filter. Google states that this scenario should not pose a major issue.

Google's algorithm uses temporal and authority signals to determine who the legitimate author is. Indexing history, content age, link profile, user behavior: these are all criteria that theoretically help distinguish the original from the copy. In this specific case, the relaunched site would retain its original status even in the face of parasitic duplicates.

Why doesn't Google systematically penalize duplicate content?

Contrary to a persistent misconception, duplicate content is not a penalty in the strict sense. Google simply filters out redundant versions to display only one URL in the results. The question is: which URL?

In the case of massive scraping, Google relies on trust signals: domain age, natural backlink profile, editorial consistency, brand signals. A legitimate site that republishes its archived content normally possesses these signals, whereas an opportunistic scraper does not have them. That’s the theory, at least.

What should you do if Google makes a mistake and favors copies?

Google recommends using the Spam Report to report sites that have siphoned your content. This is the only corrective action mentioned here, which speaks volumes about the limits of automation: the algorithm can make mistakes, especially if the authority signals of the legitimate site have been weakened by prolonged inactivity.

There isn’t an automatic mechanism to claim the precedence of content. No interdomain canonical tag, no universal authorship declaration. Manual reporting remains the only lever — and its effectiveness is never guaranteed or measurable.

  • The algorithm theoretically favors the original author through temporal and authority signals
  • Duplicate content is not a penalty but a deduplication filter in the SERPs
  • The Spam Report is the only recourse if Google preferentially indexes the copies
  • No technical mechanism to claim precedence exists from the webmaster side
  • Prolonged inactivity weakens authority signals, which can complicate automatic recognition

SEO Expert opinion

Does this statement truly reflect what we observe in the field?

Let's be honest: Google is right most of the time. In the majority of cases, an established domain that resumes its activities does not suffer a duplicate penalty against second-tier scrapers. But that "probably" in the statement is telling. There are situations where the algorithm hesitates or makes mistakes, particularly when the original site has lost its trust signals during inactivity.

A domain that has been offline for a year may see its backlinks becoming obsolete, its traffic profile disappearing, and its reputation eroding. If, in the meantime, scrapers have built an artificial link ecosystem around their copies, Google may temporarily favor those versions — especially if they generate traffic or engagement signals, even if they are false. [To be verified]: Google never discloses the precise thresholds that trigger a reversal of priority between original and copy.

What are the concrete limits of this algorithmic protection?

The problem intensifies in three specific scenarios. First case: content is republished on a new domain without history. Even if it's the same owner, Google has no technical means to automatically establish the link between the old and new domain. The signals reset to zero.

Second case: scrapers have massively enriched or modified the copied content. If the copy contains substantial additions, Google may consider it an improved version and favor it. Third case: the legitimate site generates no engagement signals upon restart (zero traffic, zero fresh backlinks), while the copies artificially generate them via PBNs or bot traffic. The algorithm may be temporarily deceived.

How effective is the Spam Report as a solution?

Google's advice — "report spam sites" — is technically correct but operationally weak. The Spam Report works on a large scale for obvious content farms, but for isolated cases of targeted scraping, the effectiveness is anecdotal. No SLA, no feedback, no guarantee of processing.

In practice, waiting for a manual action from Google is not a strategy. It's better to focus your efforts on rapidly rebuilding positive signals: qualified traffic, fresh editorial backlinks, social engagement, brand mentions. This is what convinces the algorithm, not a reporting form. [To be verified]: Google never publishes statistics on processing rates or the real impact of Spam Reports submitted by individual webmasters.

Attention: If you relaunch a site after long inactivity on a new domain, don't rely solely on Google’s automatic detection. Implement an active re-legitimation strategy: 301 redirects if possible, targeted link-building campaigns, reactivating social accounts, strong brand signals. The algorithm needs tangible proof of continuity.

Practical impact and recommendations

What should you do before relaunching an inactive site?

Before any republication, audit the extent of scraping of your content. Use tools like Copyscape, Ahrefs Content Explorer, or a simple Google search with exact excerpts in quotes. Identify how many domains have copied your pages and assess their respective authority. If you detect copies on high-authority sites, the risk of algorithmic confusion increases.

Next, prepare for the reactivation of your authority signals. If you are relaunching on the same domain, ensure that the old URLs are restored exactly to benefit from indexing history. If migrating to a new domain, set up 301 redirects from the old one — even if Google says it's not mandatory, it's a strong signal of continuity.

How can you maximize your chances of being recognized as a legitimate source?

Publish fresh content immediately after reactivation. A site that restarts with only dated content appears suspicious. Add update dates, new sections, and editorial enrichments. This proves that the site is actively maintained and not just a static clone.

Launch a mini-campaign of targeted editorial link-building in the initial weeks. A few mentions in industry media, a press release announcing the relaunch, strategic guest posts: these are all signals that enhance your legitimacy. Google places enormous weight on fresh backlinks in its authority and precedence calculations.

What mistakes should you absolutely avoid?

Never republish your archived content without any modification. Even minor adjustments — updating dates, adding sections, optimizing tags — send a signal of freshness. Content that is strictly identical to existing copies may raise doubts in the algorithm.

Also, avoid neglecting Search Console. Actively submit your sitemap, monitor crawling errors, and check that indexing is progressing normally. If you notice that copies are indexed but not your original pages, it's time to fill out Spam Reports — but without expecting miracles, as we have seen.

  • Audit the extent of scraping with Copyscape or Ahrefs Content Explorer
  • Restore the original URLs exactly or set up clean 301 redirects
  • Publish fresh content upon reactivation to prove editorial activity
  • Launch a targeted link-building campaign within the first 30 days
  • Update dates and enrich archived content before republication
  • Submit the sitemap in Search Console and actively monitor indexing
Relaunching a site after long inactivity requires a proactive approach to maximize your chances of being recognized as a legitimate source by Google. The algorithm generally makes the right choice, but it needs clear signals: technical continuity via redirects, editorial freshness via updates, regained authority through strategic link-building. These optimizations can be complex to orchestrate alone, especially when managing technical migration, content production, and link acquisition simultaneously. In this context, relying on a specialized SEO agency can ensure a methodical reactivation and prevent mistakes that might compromise the algorithmic recognition of your legitimacy.

❓ Frequently Asked Questions

Google pénalise-t-il vraiment le duplicate content ?
Non, le duplicate content n'est pas une pénalité mais un filtre de dédoublonnage. Google choisit simplement quelle version afficher dans les résultats, généralement en privilégiant l'original détecté via des signaux d'autorité et d'ancienneté.
Comment Google identifie-t-il l'auteur original d'un contenu ?
Google utilise des signaux temporels (première indexation, historique d'archives), d'autorité (profil de backlinks, confiance du domaine) et comportementaux (trafic, engagement) pour déterminer la source légitime. Ces signaux ne sont jamais détaillés publiquement.
Faut-il obligatoirement garder le même domaine pour éviter les problèmes de duplicate ?
Non, mais migrer sur un nouveau domaine efface votre historique d'indexation et affaiblit vos signaux d'autorité. Des redirections 301 depuis l'ancien domaine et une stratégie de re-légitimation active sont alors indispensables.
Le Spam Report est-il efficace contre les scrapers ?
Son efficacité est limitée pour des cas isolés. Google traite en priorité les fermes de contenu à grande échelle. Pour un scraping ciblé, mieux vaut renforcer vos propres signaux d'autorité que d'attendre une action manuelle hypothétique.
Combien de temps faut-il pour que Google reconnaisse le site légitime après un relaunch ?
Cela dépend de la vitesse à laquelle vous reconstituez vos signaux d'autorité. Avec du contenu frais, du netlinking actif et un bon suivi Search Console, quelques semaines suffisent généralement. Sans ces signaux, la reconnaissance peut prendre plusieurs mois ou ne jamais se faire correctement.
🏷 Related Topics
Algorithms Content AI & SEO JavaScript & Technical SEO Domain Name Penalties & Spam

🎥 From the same video 43

Other SEO insights extracted from this same Google Search Central video · duration 1h14 · published on 04/06/2020

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.