What does Google say about SEO? /

Official statement

Google does not have CWV data for every individual page. Pages are grouped according to the available data: at the domain/origin level if there is little data, or by groups of similar pages if there is more data. Noindex pages may be included in the Chrome UX Report because it is difficult to determine their indexability from Chrome's side.
47:03
🎥 Source video

Extracted from a Google Search Central video

⏱ 1h01 💬 EN 📅 05/02/2021 ✂ 48 statements
Watch on YouTube (47:03) →
Other statements from this video 47
  1. 2:42 Does Google penalize dynamic content on e-commerce pages?
  2. 2:42 Does variable content on e-commerce pages harm SEO?
  3. 4:15 Is Google really penalizing wide or inconsistent e-commerce categories?
  4. 4:15 Is it true that Google penalizes category pages lacking strict thematic consistency?
  5. 6:24 How does Google determine the order of images on a single page?
  6. 6:24 Does Google prioritize image quality over the display order on the page?
  7. 8:00 Is machine learning for images truly a secondary SEO factor?
  8. 8:29 Can machine learning really replace text for SEO-ing your images?
  9. 11:07 Why does Google Discover traffic seem to vanish overnight?
  10. 11:07 Why does Google Discover traffic drop off overnight without warning?
  11. 13:13 Do Google penalties really work page by page without fixed levels?
  12. 13:13 Does Google really impose page-by-page granular penalties instead of site-wide ones?
  13. 15:21 Could Google hide one of your sites if they look too similar?
  14. 15:21 Why does Google omit certain unique sites in its results?
  15. 17:29 Can a low-quality page really taint your entire site?
  16. 17:29 Can a poorly optimized homepage really penalize an entire site?
  17. 18:33 How does Google measure Core Web Vitals on your AMP and non-AMP pages?
  18. 18:33 Does Google really track Core Web Vitals for AMP and non-AMP pages separately?
  19. 20:40 Core Web Vitals: Which version truly impacts your ranking when Google shows the AMP?
  20. 22:18 Should you really match the query in the title to rank well?
  21. 22:18 Should you choose an exact match title or a user-optimized title?
  22. 24:28 Do user comments really influence your page rankings?
  23. 24:28 Do user comments really count for SEO?
  24. 28:00 Are intrusive interstitials really a negative ranking factor?
  25. 28:09 Can intrusive interstitials really lower your Google ranking?
  26. 29:09 Why does Google convert your SVGs to PNGs and how does it affect your image SEO?
  27. 29:43 Why does Google convert your SVGs into pixel images internally?
  28. 31:18 Should you optimize the user experience before tackling SEO?
  29. 31:44 Should you really use rel=canonical for syndicated content?
  30. 32:24 Does rel=canonical to the source really protect syndicated content?
  31. 34:29 Should you create broad topical content to boost your authority in Google's eyes?
  32. 34:29 Should you create related content to boost your topical authority?
  33. 36:01 How long should you really expect to wait for a manual link action to be lifted?
  34. 36:01 Why can manual link actions take several months to get a response?
  35. 39:12 Does PageSpeed Insights really reflect what Google sees on your site?
  36. 39:44 Why do PageSpeed Insights and Googlebot show different results for your site?
  37. 41:20 Is it true that your PageSpeed Insights tests don't accurately reflect what Google really measures regarding Core Web Vitals?
  38. 44:59 Do you really need to wait 30 days to see the impact of your Core Web Vitals optimizations in PageSpeed Insights?
  39. 45:59 Core Web Vitals: Why Do Only Real User Data Matter for Ranking?
  40. 45:59 Why does Google overlook your Lighthouse scores when ranking your site?
  41. 46:43 How does Google really group your pages to evaluate Core Web Vitals?
  42. 51:24 Why does Google keep crawling outdated 404 URLs on your site?
  43. 51:54 Why does Google keep rechecking your old 404 URLs for years?
  44. 57:06 Do 301 redirects really pass on 100% of PageRank and link signals?
  45. 57:06 Do 301 redirects really transfer all ranking signals without any loss?
  46. 59:51 Is it true that the text/HTML ratio is completely irrelevant for Google SEO?
  47. 59:51 Is the text/HTML ratio really useless for SEO?
📅
Official statement from (5 years ago)
TL;DR

Google does not measure Core Web Vitals on a page-by-page basis. The granularity depends on the volume of data: grouping at the domain level for small sites, by clusters of similar pages for larger ones. Noindex pages can appear in CrUX because Chrome does not necessarily detect their indexability status. The result: the metrics displayed in the Search Console do not always reflect the reality of each URL.

What you need to understand

Why doesn’t Google have CWV data for each individual page?

The Chrome User Experience Report (CrUX) collects real performance data from Chrome browsers, not from Googlebot. This collection depends on the volume of actual visits: a minimally visited page does not generate enough data to produce a reliable and representative metric.

Thus, Google applies an adaptive grouping mechanism. For low-traffic sites, metrics are aggregated at the domain level. For sites receiving more traffic, Google creates groups of similar pages — for example, all product pages, all category pages — and calculates Core Web Vitals at the scale of these clusters.

How does this affect the interpretation of metrics in the Search Console?

The reports displayed in the Search Console present aggregated averages, not URL-by-URL measurements. An individual page may perform well even while being lumped into a group marked as 'Poor' due to structural issues affecting other pages of the same template.

Conversely, a problematic page may go unnoticed if the overall group shows good performance. This logic necessitates reasoning by type of page rather than by isolated URL when auditing and fixing CWV issues.

Why do noindex pages appear in CrUX?

Chrome collects performance data regardless of indexability. The browser does not parse meta robots tags, does not check robots.txt, and does not verify X-Robots-Tag headers. It simply measures user experience across all pages visited by real users.

A direct consequence: pages blocked from indexing, poorly secured staging URLs, or test pages in production can pollute your CrUX metrics if they receive traffic. Google aggregates this data into the corresponding groups, potentially skewing the analysis for SEOs who expect to see only indexable pages.

  • CrUX measures real user experience, not a page's ability to be indexed
  • Metrics are grouped by page similarity or at the domain level depending on the volume of data
  • A noindex page can influence your Core Web Vitals if it receives organic, direct, or referral traffic
  • Search Console reports display group averages, rarely page-by-page data
  • The granularity of data depends on traffic: the more visited the site is, the finer the groups can become

SEO Expert opinion

Is this grouping mechanism consistent with real-world observations?

Yes, absolutely. SEO audits have confirmed for years that the CWV reports from the Search Console present aggregated data by template or page type. An e-commerce site often sees its product pages grouped together, its category pages in another cluster, and its homepage isolated if it receives significant traffic.

What still surprises many practitioners is the latency between fixing and updating metrics. Fixing a Cumulative Layout Shift issue on one page doesn’t immediately change anything in the Search Console: we must wait for the group to accumulate enough new real measurements, which can take several weeks. [To be verified]: Google has never specified the minimum data threshold required to create a distinct group.

What are the grey areas and limitations of this statement?

Mueller remains deliberately vague on the grouping criteria. We know that Google uses 'similar pages', but no official documentation details the criteria precisely: common HTML structure? URLs sharing a pattern? Identical backend template? This opacity makes blind optimization difficult for sites with complex architecture.

Another critical point: the presence of noindex pages in CrUX creates a methodological paradox. Google recommends not indexing certain pages to avoid duplicate content or low-value pages, but these same pages can degrade the Core Web Vitals of the group they belong to. The result: a site can be penalized on a ranking criterion because of pages explicitly excluded from ranking.

When does this grouping system pose problems?

Sites with hybrid architectures are most affected. Imagine a media site that mixes lightweight AMP articles, long-format resource-heavy reports, and event pages with heavy third-party scripts. If Google groups everything into a cluster 'editorial content', the metrics become unusable to diagnose precisely where to act.

International sites with language versions on the same domain also encounter grouping ambiguities. Does CrUX aggregate by language, by user geolocation, or globally? No official confirmation. The same applies to sites with separate URLs for mobile and desktop versions: are the data separated or merged?

Attention: If your site shows significant discrepancies between CrUX metrics (real data) and Lighthouse tests (lab data), it is likely that the grouping includes problematic pages you do not identify during your manual audits. Dig deeper into real traffic by page type to isolate impacted clusters.

Practical impact and recommendations

What should you prioritize in an audit to understand your CWV groupings?

Start by exporting CrUX data via BigQuery or the CrUX API. Unlike the Search Console, these sources offer URL-by-URL granularity for pages with sufficient traffic. Cross-reference this data with your Analytics tagging plan to identify which templates or site sections are dragging down the metrics.

Next, list all noindex or blocked pages that receive significant real traffic — old URLs soft-404 redirected, preview pages accessible without authentication, poorly isolated testing environments. If these pages correspond to a template also used in indexable production, they pollute your CWV groups.

How to optimize when metrics are grouped rather than individual?

Think by type of page, not by URL. Identify common templates — product page, category page, blog post, landing page — and optimize the source code at the template level. A 200 ms gain on the Largest Contentful Paint of a product page template benefits the entire group, not just one isolated page.

Prioritize templates that generate the most qualified organic traffic. A template used by 10,000 pages but with a total of 50 monthly visits does not impact Core Web Vitals as much as a template used by 100 pages receiving 50,000 visits. Focus your efforts where the volume of real data is sufficient to influence the grouping.

What mistakes should you avoid in tracking and correcting Core Web Vitals?

Never rely solely on Lighthouse or PageSpeed Insights to validate your optimizations. These tools measure in a lab environment, with standardized network and hardware conditions. The Core Web Vitals that count for ranking come from CrUX, thus from real users with variable connections, heterogeneous devices, and unpredictable browsing behaviors.

Also, avoid fixing page by page without a comprehensive view. If you manually optimize 20 critical URLs but the common template used by 500 other pages remains faulty, the grouping will continue to show poor metrics. The effort must be structural: template redesign, lazy loading at the CMS level, global server optimization, CDN for static resources.

  • Export CrUX data via BigQuery or the API to identify problematic URLs and groups
  • List all noindex pages receiving traffic and check if they pollute their group's metrics
  • Prioritize optimizations by template and real traffic volume, not by subjective editorial importance
  • Cross-reference CrUX data with Analytics to understand which user journeys degrade metrics
  • Test corrections in production with Real User Monitoring (RUM) tools before validating their effectiveness
  • Monitor group evolution in the Search Console for at least 28 days to confirm the impact of optimizations
The adaptive grouping of Core Web Vitals requires thinking about optimization in a systemic and architectural manner, not URL by URL. Complex sites with heterogeneous templates, noindex pages receiving traffic, or international architecture require a thorough audit to isolate the real improvement levers. These technical diagnostics and the resulting strategic decisions are often complex to manage internally without dedicated expertise. For high-stakes sites, being assisted by an SEO agency specialized in technical optimization and Core Web Vitals can significantly accelerate achieving 'Good' thresholds and secure the associated ranking gains.

❓ Frequently Asked Questions

Est-ce que toutes mes pages ont des données Core Web Vitals individuelles dans la Search Console ?
Non. Google regroupe les pages par similarité ou au niveau du domaine si le trafic est insuffisant. Seules les pages avec un volume élevé de données réelles peuvent afficher des métriques individuelles.
Pourquoi mes pages noindex apparaissent-elles dans le rapport CrUX ?
Le CrUX collecte les données depuis Chrome, qui ne vérifie pas l'indexabilité. Toute page visitée par un utilisateur réel génère des métriques, qu'elle soit indexable ou non.
Comment savoir dans quel groupe mes pages sont regroupées ?
Google ne détaille pas explicitement les groupes dans la Search Console. Utilisez l'API CrUX ou BigQuery pour obtenir des données URL par URL quand disponibles, puis croisez avec votre architecture de templates.
Un petit site a-t-il des Core Web Vitals regroupés au niveau du domaine entier ?
Oui, si le trafic est trop faible pour segmenter par groupes de pages. Toutes les pages contribuent alors à une métrique unique, ce qui rend l'optimisation ciblée plus difficile.
Combien de temps faut-il pour que mes corrections CWV apparaissent dans la Search Console ?
Les données CrUX sont basées sur des mesures réelles accumulées sur 28 jours glissants. Il faut donc attendre plusieurs semaines après correction pour voir l'impact réel, selon le volume de trafic.
🏷 Related Topics
Domain Age & History Crawl & Indexing AI & SEO JavaScript & Technical SEO Domain Name Web Performance

🎥 From the same video 47

Other SEO insights extracted from this same Google Search Central video · duration 1h01 · published on 05/02/2021

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.