Official statement
Other statements from this video 12 ▾
- 0:32 Le service de rendu Google bloque-t-il vos ressources cross-origin à cause de CORS ?
- 1:03 Les données dupliquées dans vos balises script pénalisent-elles vraiment votre SEO ?
- 1:03 La lazy hydration peut-elle vraiment tuer votre crawl budget ?
- 2:08 Pourquoi Google ne peut-il pas partager le cache JavaScript entre vos domaines ?
- 2:41 Google sur-cache-t-il vraiment les ressources de votre site ?
- 4:14 Le cache JavaScript de Google fonctionne-t-il vraiment par origine et non par domaine ?
- 6:46 Pourquoi les outils de test Google ne reflètent-ils jamais ce que voit vraiment Googlebot ?
- 7:12 Pourquoi Google ignore-t-il vos images lors du rendu pour l'indexation ?
- 12:28 Pourquoi Google insiste-t-il sur les media queries plutôt que le user-agent pour le responsive ?
- 15:16 Les outils de test Google donnent-ils vraiment les mêmes résultats ?
- 20:05 Les erreurs serveur intermittentes impactent-elles vraiment votre indexation Google ?
- 21:03 Google peut-il vraiment détecter les erreurs de rendu JavaScript sur mon site ?
Martin Splitt recommends using the 'crawled page' version of the URL Inspection Tool rather than the 'live test' to understand how Googlebot truly sees your page. The difference? The crawled test reflects cached indexing, while the live test queries your server directly. In practical terms, if you're diagnosing an indexing problem without checking the correct version, you might end up on a wild goose chase.
What you need to understand
What is the real difference between 'crawled page' and 'live test'?
The URL Inspection Tool in Search Console offers two viewing modes: the crawled version and the live test. The live test queries your server at that exact moment, without going through Google's cache. It retrieves the page as it is currently available on your infrastructure.
The crawled version, however, reflects what Googlebot actually saw during its last visit. It uses cached resources — CSS, JavaScript, images — which can create discrepancies with the current reality of your site. And it's precisely this discrepancy that matters for understanding why a page isn't indexed as you hoped.
Why does caching make a difference for indexing?
Googlebot does not load all resources in real-time at every visit. It relies on a caching system to optimize its crawl budget and processing speed. If your critical CSS has been cached for three weeks and you made changes yesterday, the live test will show the new version, but the crawled version will display the old one.
In practical terms? You might diagnose a JavaScript rendering issue that no longer exists in production, or conversely, miss a bug that Googlebot still sees. Sites that frequently deploy or rely on third-party CDNs are particularly prone to these discrepancies.
In what scenarios does this distinction become critical?
Let's take a classic case: you fix a noindex or canonicalization issue. The live test confirms that everything is OK. You validate and move on. Yet three days later, the page is still not indexed. Why? Because the crawled version is still using the old cached meta robots tag.
Another frequent scenario: JavaScript-heavy sites that load content asynchronously. The live test may display a complete rendering because your connection and server respond quickly. But the crawled version reveals that Googlebot timed out on a third-party script and never saw your main content. The result: partial or no indexing.
- The live test is used to check if a fix works before the next crawl
- The crawled version shows the actual state of indexing and caching issues
- Differences between the two often reveal performance problems or third-party resource issues
- For diagnosing an existing indexing problem, starting from the crawled version is the only reliable approach
- Sites with CDNs, external scripts, or frequent deployments should systematically compare both versions
SEO Expert opinion
Does this recommendation align with observed practices in the field?
Yes, and it's even a recurring friction point in SEO audits. Practitioners heavily use the live test because it feels immediate and gives the impression of testing 'the real page'. Except that indexing doesn't work that way. Google indexes what it has crawled, not what is theoretically available on your server at that precise moment.
We regularly observe cases where a client claims that 'everything is fixed' based on the live test, while the crawled version still shows the issue. The lag between the two can range from a few hours to several weeks depending on the crawl budget allocated to the site. For low-authority sites or deep pages, this discrepancy becomes a classic trap.
What nuances should be added to this statement?
First point: the crawled version is reliable only if it is recent. If Googlebot hasn't crawled in three months, this version won't tell you anything about your current indexability status. In this case, the live test remains your only tool to validate a hypothesis — but you'll need to force a new crawl to check the real impact.
Second nuance: the live test retains utility for A/B tests or changes in the process of deployment. Want to know if your new HTML structure is crawlable before deploying it to production? The live test is perfect for that. But once in production, return to the crawled version for post-mortem analysis. [To be verified]: Google does not document precisely how long resources remain cached or how this timeframe is calculated. We know it varies by resource type, but the exact metrics remain opaque.
In what cases does this rule not apply or pose problems?
If you're managing a news site or an e-commerce platform with hourly updates, the crawled version will almost always be outdated. In this context, the live test becomes your reference for validating the structure, and you need to monitor server logs to understand what Googlebot actually sees during its frequent visits.
Another edge case: sites with server-side personalization or A/B testing. The crawled version reflects a potentially different variant than the one you're testing live. You risk diagnosing an issue on a variant that Googlebot has never seen, or validating a fix on a version that will never be crawled. Here, you need to cross-reference the two tools with logs to triangulate reality.
Practical impact and recommendations
What should you do concretely to diagnose an indexing issue?
Always start with the crawled version in the URL Inspection Tool. Check the date of the last crawl — if it dates back more than a week and you have made recent changes, request reindexing and come back a few days later. Do not draw any conclusions from an outdated version.
Next, compare both versions. If the live test shows a correct rendering but not the crawled version, you likely have a caching issue, JavaScript timeout, or third-party resource problems. Download the raw HTML of both versions and do a diff to identify precisely what diverges. Tools like Beyond Compare or WinMerge are perfect for this.
What mistakes should be avoided during the analysis?
Classic error: relying solely on the rendering screenshot without checking the source HTML. Googlebot may display a visually correct rendering but have missed critical elements (hreflang, structured data, canonical) because they are injected too late by JavaScript. Scroll to the source code and check manually.
Another trap: ignoring blocked resources. Search Console flags blocked CSS or JS at the bottom of the inspection. If a critical resource is blocked in the crawled version but not in the live test, you likely have a robots.txt issue that was recently fixed but whose cache has not expired. Force a new crawl of the robots.txt via Search Console.
How to integrate this practice into an SEO audit workflow?
Integrate a systematic crawled validation step after every structural change. Don't just validate in staging or with the live test — wait for the next production crawl and check that Google sees what you expect. For critical sites, set up monitoring with the Search Console API to automatically track discrepancies between live and crawled.
For complex projects with client-side JavaScript, combine this approach with a headless crawler (Screaming Frog in Chrome mode, Sitebulb, or a custom Puppeteer script) to simulate Googlebot rendering independently. Triangulating the three sources — GSC crawled version, live test, third-party crawler — gives you a reliable view of reality.
- Always check the date of the last crawl before drawing conclusions
- Compare the source HTML, not just the rendering screenshot
- Download and diff the two versions to identify precise divergences
- Check for blocked resources and loading errors in the crawled version
- Force reindexing after correction and wait for the new crawl to validate
- Set up monitoring via the Search Console API for critical sites
❓ Frequently Asked Questions
Pourquoi le live test et la version crawlée diffèrent-ils ?
Dans quels cas le live test reste-t-il utile ?
La version crawlée peut-elle être obsolète ?
Comment interpréter des différences de rendu entre les deux versions ?
Cette recommandation s'applique-t-elle aussi pour les sites JavaScript ?
🎥 From the same video 12
Other SEO insights extracted from this same Google Search Central video · duration 26 min · published on 15/10/2020
🎥 Watch the full video on YouTube →
💬 Comments (0)
Be the first to comment.