What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 3 questions

Less than 30 seconds. Find out how much you really know about Google search.

🕒 ~30s 🎯 3 questions 📚 SEO Google

Official statement

When a page lacks sufficient CrUX data, Google uses the scores of similar pages from the same site to make an estimation. If the site's structure is complex, the overall site score may be applied to the concerned page.
22:22
🎥 Source video

Extracted from a Google Search Central video

⏱ 1h07 💬 EN 📅 28/01/2021 ✂ 28 statements
Watch on YouTube (22:22) →
Other statements from this video 27
  1. 13:31 Vos pages lentes peuvent-elles plomber le classement de tout votre site ?
  2. 13:33 Les Core Web Vitals impactent-ils vraiment tout votre site ou seulement vos pages lentes ?
  3. 13:33 Peut-on bloquer la collecte des Core Web Vitals avec robots.txt ou noindex ?
  4. 14:54 Pourquoi CrUX collecte vos Core Web Vitals même si vous bloquez Googlebot ?
  5. 15:50 Page Experience : Google ment-il sur son véritable poids dans le classement ?
  6. 16:36 L'expérience de page est-elle vraiment un signal de classement secondaire ?
  7. 17:28 Le LCP mesure-t-il vraiment la vitesse perçue par l'utilisateur ?
  8. 19:57 Les Core Web Vitals se calculent-ils vraiment pendant toute la navigation ?
  9. 20:04 Les Core Web Vitals évoluent-ils vraiment après le chargement initial de la page ?
  10. 21:22 Comment Google estime-t-il vos Core Web Vitals quand les données CrUX manquent ?
  11. 27:07 Comment Google attribue-t-il désormais les données CrUX du cache AMP à l'origine ?
  12. 29:47 AMP est-il encore nécessaire pour ranker dans Top Stories sur mobile ?
  13. 32:31 Comment exploiter les logs serveur pour détecter les erreurs 4xx dans Search Console ?
  14. 34:34 Pourquoi les nouveaux sites connaissent-ils une volatilité extrême dans l'indexation et le classement ?
  15. 34:34 Faut-il vraiment analyser les logs serveur pour diagnostiquer les erreurs 4xx dans Search Console ?
  16. 34:34 Pourquoi votre nouveau site fluctue-t-il comme un yoyo dans les SERP ?
  17. 40:03 Faut-il vraiment signaler le contenu copié de votre site via le formulaire spam de Google ?
  18. 40:20 Comment signaler efficacement le spam de contenu copié à Google ?
  19. 43:43 Vos pages franchise sont-elles des doorway pages aux yeux de Google ?
  20. 45:46 Le contenu dupliqué est-il vraiment sans danger pour votre référencement ?
  21. 45:46 Le contenu dupliqué est-il vraiment sans pénalité pour votre SEO ?
  22. 45:46 Vos pages franchises sont-elles perçues comme des doorway pages par Google ?
  23. 51:52 Le namespace http:// ou https:// dans un sitemap XML influence-t-il vraiment le crawl ?
  24. 52:00 Le namespace en https dans votre sitemap XML pénalise-t-il votre référencement ?
  25. 55:56 Faut-il vraiment inclure les deux versions mobile et desktop dans son sitemap XML ?
  26. 56:00 Faut-il vraiment soumettre les versions mobile ET desktop dans votre sitemap ?
  27. 61:54 Faut-il abandonner AMP si vous utilisez GA4 pour mesurer vos performances ?
📅
Official statement from (5 years ago)
TL;DR

Google uses scores from similar pages on the same site to estimate the Core Web Vitals of a page lacking CrUX data. This estimation logic—based on proximity or a default site-wide score—directly impacts the ranking of under-visited pages. For an SEO, this means that optimizing high-traffic pages can indirectly enhance the overall domain score.

What you need to understand

What is CrUX and why do some pages lack data?

The Chrome User Experience Report (CrUX) aggregates real user performance metrics collected from Chrome users. For a page to have sufficient CrUX data, it must receive a significant volume of visits over a given period.

Pages with low organic traffic (deep pages, niche content, new pages) often do not meet this threshold. Google then has no reliable metrics to evaluate their Core Web Vitals. Specifically, a little-visited product page or an old blog page with few visitors will not have its own CrUX score.

How does the 'similar site' estimation work?

Google employs a fallback logic: if a page lacks data, the algorithm first looks for 'similar' pages within the same domain. Similarity likely depends on HTML structure, common templates, content type, or navigation depth.

If the site's structure is too complex or heterogeneous, Google defaults to the overall domain score. In other words, the average rating of all CWV for the site becomes the proxy for this isolated page. This approach prevents penalizing an entire site for a handful of pages without data, but it also obscures local performance variations.

Why is this statement important for an SEO?

It confirms that Google does not rely solely on direct measures for all content. Therefore, a site can improve the perceived score of hundreds of pages by optimizing only those that generate traffic—and thus measurable CrUX data.

Conversely, a site with a few slow yet highly visited pages risks contaminating the score of the entire domain. This estimation logic creates a side effect: optimizing poor performers can disproportionately impact the overall ranking.

  • Pages without CrUX data inherit the score from similar pages or the entire domain
  • The site's structure influences the choice of estimation (similarity or global)
  • Optimizing high-traffic pages can improve the score of invisible pages
  • A heterogeneous site (varied templates, disparate performances) risks more fallback to the global score
  • This logic is opaque: Google does not detail the 'similarity' criteria

SEO Expert opinion

Is this statement consistent with field observations?

Yes, it aligns with feedback from SEOs who have observed ranking improvements on lesser-visited pages after optimizing main templates. If Google were relying solely on PageSpeed Insights (synthetic tests), these variations would be inexplicable. The effect of CrUX score propagation confirms this mechanism.

However, Google remains deliberately vague about the criteria for 'similar pages'. It’s believed to involve DOM patterns, common CSS classes, similar load times—but there’s no confirmation. [To verify]: sites with multiple templates (blog, e-commerce, landing pages) could experience inconsistent estimations if the algorithm does not segment correctly.

What are the limitations and grey areas of this approach?

The statement does not specify the threshold of sufficient data. Does a page with 100 visits per month have usable CrUX data? 500? 1000? This opacity makes it difficult to anticipate exactly which pages will be estimated and which will have their own score.

Another point: what happens if a 'similar' page itself has an estimated score? Does Google apply a cascading estimation or switch directly to the global score? This question remains unresolved. Furthermore, the notion of 'complex structure' is subjective—does a site with 10 different templates qualify as 'complex'? [To verify].

What is the real risk for an e-commerce or media site?

An e-commerce site with thousands of little-visited product pages may see the majority of its pages estimated based on the global score or a handful of prototype pages. If the homepage and main categories are slow, the entire catalog inherits this penalty, even if the product pages are technically light.

For media sites, the risk is the opposite: highly consulted news articles (often overloaded with ad scripts) can drag down the score of the entire archive. Let's be honest: this logic favors homogeneous and well-structured sites—and penalizes technical patchworks accumulated over the years.

Attention: An isolated CWV audit (PageSpeed Insights) does not necessarily reflect the CrUX score used by Google for ranking. A page can show good synthetic results but inherit a poor overall domain score.

Practical impact and recommendations

What should you prioritize auditing on your site?

Start by identifying the pages that generate real traffic and thus measurable CrUX data. Google Search Console, combined with CrUX via PageSpeed Insights API, allows you to cross-reference URLs with organic traffic and available CWV scores.

Next, segment your site by templates or content types: homepage, categories, product pages, blog posts, conversion pages. For each segment, measure CWV under real conditions (RUM, CrUX Dashboard). If a high-volume template shows poor scores, it potentially contaminates the entire estimated segment.

How can you optimize to reduce the negative impact of estimations?

Favor a homogeneous architecture: the fewer different templates you have, the more accurately Google can refine its similarity estimations. Unify third-party scripts, critical CSS, and loading patterns as much as possible. A site with 3 well-optimized templates is better than a site with 15 disparate templates.

For high-traffic pages, apply classic optimizations: aggressive lazy loading, AVIF/WebP compression, font preloading, elimination of blocking JavaScript, efficient CDN. These pages serve as a reference for the estimation—taking care of them equates to caring for the entire domain. And that’s where it gets tricky: if your infrastructure is aging or technical teams do not prioritize CWV, these optimizations can remain in the backlog for months.

What mistakes should you absolutely avoid?

Do not focus solely on strategic pages (conversion, top 3 ranking). A little-visited page but belonging to a large segment can inherit the average score of the segment. If all your product pages are slow except for 10, the 10 fast ones will not be sufficient to compensate.

Avoid multiplying custom templates for every type of page. The more you fragment, the more likely Google is to default to the global domain score. Lastly, do not neglect pages without organic traffic but accessible (sitemap, internal linking): they are indexed, estimated, and can influence the site's overall perception. Specifically? Regularly clean up obsolete or poorly performing pages that pollute your average score.

  • Identify pages with real vs estimated CrUX data (CrUX API + Search Console)
  • Segment the site by templates and measure the average CWV by segment
  • Prioritize optimizing high-traffic templates
  • Homogenize the technical architecture (scripts, CSS, loading patterns)
  • Remove or noindex obsolete pages with poor CWV that drag down the global score
  • Monitor the evolution of CrUX scores at the domain level (CrUX Dashboard BigQuery)
Google's estimation of Core Web Vitals relies on an opaque but predictable fallback logic: similar pages first, global domain score next. For an SEO, this means that optimizing a few dozen strategic pages can enhance the ranking of hundreds of lesser-visited pages. Conversely, neglecting main templates contaminates the entire site. This mechanism favors homogeneous architectures and penalizes patchwork sites. If your infrastructure is complex or if technical resources are limited, these optimizations can quickly become time-consuming and require specialized support. An experienced SEO agency can help you prioritize tasks, finely audit page segments, and guide development with technical teams—often essential for unlocking sustainable gains on high-volume sites.

❓ Frequently Asked Questions

Comment savoir si ma page utilise des données CrUX réelles ou estimées ?
Utilise l'API CrUX ou le CrUX Dashboard (BigQuery). Si une URL n'apparaît pas dans le dataset, elle est estimée par similarité ou via le score global du domaine. PageSpeed Insights affiche parfois un avertissement si les données sont insuffisantes.
Le score PageSpeed Insights reflète-t-il le score CrUX utilisé pour le ranking ?
Pas toujours. PageSpeed Insights affiche un test synthétique (Lighthouse) et, si disponibles, les données CrUX réelles. Mais une page peut hériter d'un score CrUX estimé (via pages similaires ou domaine global) qui diffère du test synthétique. Seul le CrUX compte pour le ranking.
Est-ce que toutes les pages d'un même template ont le même score CrUX estimé ?
C'est probable si Google les considère comme « similaires », mais la déclaration reste floue sur les critères exacts. Des pages partageant le même template mais avec des contenus très différents (images lourdes, scripts tiers variables) pourraient recevoir des estimations distinctes.
Faut-il optimiser les pages sans trafic organique pour améliorer le score global ?
Si ces pages sont indexées et nombreuses, elles influencent le score global du domaine utilisé pour les estimations. Nettoyer ou améliorer ces pages peut donc avoir un impact indirect. Mais priorise d'abord les pages à fort trafic, qui génèrent des données CrUX mesurables.
Un site avec plusieurs sous-domaines a-t-il un score CrUX global unique ?
Non, chaque sous-domaine est traité comme un site distinct dans CrUX. Si blog.example.com a de mauvais CWV, cela n'affecte pas directement le score de shop.example.com. En revanche, les estimations par similarité restent cantonnées à chaque sous-domaine.
🏷 Related Topics
Domain Age & History AI & SEO Pagination & Structure

🎥 From the same video 27

Other SEO insights extracted from this same Google Search Central video · duration 1h07 · published on 28/01/2021

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.