Official statement
Other statements from this video 9 ▾
- 1:45 Pourquoi Google n'indexe-t-il pas le contenu qu'il ne parvient pas à rendre en JavaScript ?
- 3:01 Pourquoi Google n'indexe-t-il pas toutes les pages des gros sites ?
- 5:45 Les Core Updates changent-ils vraiment le classement en continu entre deux mises à jour ?
- 9:48 Le maillage interne suffit-il vraiment à booster le classement de toutes vos pages ?
- 10:20 Les blogs rankent-ils plus vite que les pages statiques dans Google ?
- 14:37 Pourquoi Google affiche-t-il parfois des URLs M-Dot dans les résultats desktop ?
- 23:54 Les erreurs 500 prolongées font-elles vraiment disparaître vos pages de l'index Google ?
- 32:09 Faut-il vraiment utiliser l'outil de changement d'adresse pour migrer des sous-domaines ?
- 53:20 Pourquoi Google peut-il fusionner vos pages JS si les balises meta sont identiques ?
John Mueller states that a misconfiguration of the Vary header on a responsive site does not directly impact indexing but causes unnecessary recrawls by Googlebot. Specifically, your pages remain indexed, but you waste crawl budget without a valid reason. The issue is therefore less catastrophic than one might think, but it remains an optimization problem to address for high-volume sites.
What you need to understand
What is the Vary header and why does it exist?
The Vary header is an HTTP mechanism that allows servers to indicate to browsers and crawlers which variables influence the version of content served. For a responsive site that serves the same HTML to all devices, the Vary header is useless — as the same version is sent to everyone.
The problem arises when developers configure Vary: User-Agent out of habit or ignorance. This directive signals to Googlebot that the content may change depending on the User-Agent, which contradicts the very architecture of responsive design. The bot then interprets that it may need to recrawl the page with different User-Agents to ensure all versions are properly indexed.
Why does Googlebot unnecessarily recrawl these pages?
Googlebot detects the Vary: User-Agent header and assumes that the content differs based on whether it's crawling with a mobile or desktop User-Agent. In doubt, it schedules additional passes to check if the two versions present significant differences.
These recrawls do not harm indexing — the page remains indexed, the content is understood and ranked normally. However, you are consuming crawl budget unnecessarily, which is problematic for sites with tens of thousands of pages. Each unnecessary pass is a pass that doesn't happen on a URL that actually needs it.
Does this statement change anything about best practices?
No, it simply confirms what practitioners have already observed: a poor technical configuration may slow down crawling without breaking indexing. Mueller reassures that there is no direct penalty, but this does not mean the problem should be ignored.
For small sites, the impact will be marginal — Googlebot has plenty of time to crawl everything. For high-volume platforms, e-commerce, or media sites, it’s another story: each unnecessary recrawl is a missed opportunity to explore new pages or detect important updates.
- Vary: User-Agent on a responsive site is an inconsistent configuration that generates superfluous recrawls
- Indexing is not compromised, but crawl budget is wasted
- High-volume sites are the most affected by this inefficiency
- The fix is simple: remove the Vary header or configure it correctly according to the actual technical architecture
- Google does not penalize this error, but it remains a hindrance to crawl optimization
SEO Expert opinion
Is this statement consistent with field observations?
Yes, completely. Technical audits regularly show sites serving the same responsive content with a Vary: User-Agent in place. In server logs, we see that Googlebot visits these URLs more frequently, alternating between mobile and desktop User-Agent, without impacting ranking or presence in the index.
What Mueller does not explicitly state is how significant this waste can be. On a site with 50,000 URLs and a limited crawl budget, a doubling of the necessary passes can delay the discovery of new content by several days. This is measurable in Search Console data and logs, but Google remains vague about the actual extent of the problem. [To be verified] according to your volume and update frequency.
What nuances should be added to this statement?
Mueller speaks of a responsive site, but the situation differs with dynamic serving (serving different HTML according to the User-Agent). In this case, the Vary: User-Agent header is not only recommended but mandatory for Google to understand that it needs to crawl both versions.
Thus, the problem does not arise from the Vary header itself, but from its use inconsistently with the technical architecture. If you are using responsive design, do not use Vary. If you are using dynamic serving, Vary is essential. Confusion often arises from poorly managed architectural changes: the site transitions to responsive but the HTTP headers remain configured for dynamic serving.
In what cases does this rule not apply?
If you are actually serving different content based on the User-Agent — for example, a mobile subdomain with a different HTML structure — then the Vary header is relevant and the recrawls are justified. Google needs to see both versions to index them correctly.
Another edge case: some CDNs or proxies automatically add Vary headers to manage their internal cache without the site owner being aware. In this case, the problem is not solved on the origin server but in the CDN configuration. An analysis of the actual HTTP responses sent by the CDN is necessary to identify the source of the problem.
Practical impact and recommendations
How can you check if your site is sending an unnecessary Vary header?
Use your browser's DevTools (Network tab) or a tool like cURL to inspect the HTTP response headers of your main pages. If you see Vary: User-Agent while your site is responsive, you have identified the problem.
Check your server logs to see if Googlebot crawls your URLs alternating between mobile and desktop User-Agent over a short period. A systematic double crawl pattern is a good indicator that Google is reacting to this inconsistent Vary header. The Search Console will not directly alert you on this point — it's in the logs where this is visible.
What should you do concretely to correct this error?
If your site is truly responsive and serves the same HTML to all devices, remove the Vary: User-Agent header from your HTTP responses. On Apache servers, this is done in the .htaccess or vhost configuration. On Nginx, in the relevant server or location block.
If you are using a CDN (Cloudflare, Fastly, Akamai, etc.), check the cache configuration: some CDNs automatically add a Vary: User-Agent to manage their own caching policies. In this case, disable this rule or adjust the configuration so that it does not apply to crawlers. Then test with a crawler on your end to validate that the change has been properly implemented.
What errors should be avoided when making this correction?
Do not remove the Vary header if you are actually using dynamic serving — you would then break Google's ability to detect that multiple versions exist. Ensure first that your architecture is truly responsive before making changes to the headers.
Avoid modifying headers without testing the impact on the CDN cache. A sudden change may lead to a cache purge and a temporary increase in server load. Plan for a gradual rollout if you manage a high-traffic site. Finally, document the change in your technical procedures to avoid a future deployment inadvertently reintroducing the error.
- Inspect HTTP response headers to detect an unnecessary Vary: User-Agent
- Analyze server logs to identify patterns of double crawling by Googlebot
- Remove the Vary header if the site is responsive (same HTML for all devices)
- Check the CDN configuration to ensure it does not automatically add this header
- Test the change with a crawler before deployment in production
- Document the change to prevent future regressions
❓ Frequently Asked Questions
L'en-tête Vary peut-il vraiment empêcher l'indexation de mes pages ?
Comment savoir si Googlebot crawle mes pages en double à cause de Vary ?
Dois-je garder l'en-tête Vary si je fais du dynamic serving ?
Mon CDN ajoute automatiquement un Vary: User-Agent, que faire ?
Combien de temps faut-il pour que Google arrête les recrawls inutiles après correction ?
🎥 From the same video 9
Other SEO insights extracted from this same Google Search Central video · duration 56 min · published on 04/09/2019
🎥 Watch the full video on YouTube →
💬 Comments (0)
Be the first to comment.