What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 5 questions

Less than a minute. Find out how much you really know about Google search.

🕒 ~1 min 🎯 5 questions

Official statement

Incorrect use of the Vary header for responsive sites does not harm indexing but can lead to unnecessary recrawls by Googlebot.
29:06
🎥 Source video

Extracted from a Google Search Central video

⏱ 56:43 💬 EN 📅 04/09/2019 ✂ 10 statements
Watch on YouTube (29:06) →
Other statements from this video 9
  1. 1:45 Pourquoi Google n'indexe-t-il pas le contenu qu'il ne parvient pas à rendre en JavaScript ?
  2. 3:01 Pourquoi Google n'indexe-t-il pas toutes les pages des gros sites ?
  3. 5:45 Les Core Updates changent-ils vraiment le classement en continu entre deux mises à jour ?
  4. 9:48 Le maillage interne suffit-il vraiment à booster le classement de toutes vos pages ?
  5. 10:20 Les blogs rankent-ils plus vite que les pages statiques dans Google ?
  6. 14:37 Pourquoi Google affiche-t-il parfois des URLs M-Dot dans les résultats desktop ?
  7. 23:54 Les erreurs 500 prolongées font-elles vraiment disparaître vos pages de l'index Google ?
  8. 32:09 Faut-il vraiment utiliser l'outil de changement d'adresse pour migrer des sous-domaines ?
  9. 53:20 Pourquoi Google peut-il fusionner vos pages JS si les balises meta sont identiques ?
📅
Official statement from (6 years ago)
TL;DR

John Mueller states that a misconfiguration of the Vary header on a responsive site does not directly impact indexing but causes unnecessary recrawls by Googlebot. Specifically, your pages remain indexed, but you waste crawl budget without a valid reason. The issue is therefore less catastrophic than one might think, but it remains an optimization problem to address for high-volume sites.

What you need to understand

What is the Vary header and why does it exist?

The Vary header is an HTTP mechanism that allows servers to indicate to browsers and crawlers which variables influence the version of content served. For a responsive site that serves the same HTML to all devices, the Vary header is useless — as the same version is sent to everyone.

The problem arises when developers configure Vary: User-Agent out of habit or ignorance. This directive signals to Googlebot that the content may change depending on the User-Agent, which contradicts the very architecture of responsive design. The bot then interprets that it may need to recrawl the page with different User-Agents to ensure all versions are properly indexed.

Why does Googlebot unnecessarily recrawl these pages?

Googlebot detects the Vary: User-Agent header and assumes that the content differs based on whether it's crawling with a mobile or desktop User-Agent. In doubt, it schedules additional passes to check if the two versions present significant differences.

These recrawls do not harm indexing — the page remains indexed, the content is understood and ranked normally. However, you are consuming crawl budget unnecessarily, which is problematic for sites with tens of thousands of pages. Each unnecessary pass is a pass that doesn't happen on a URL that actually needs it.

Does this statement change anything about best practices?

No, it simply confirms what practitioners have already observed: a poor technical configuration may slow down crawling without breaking indexing. Mueller reassures that there is no direct penalty, but this does not mean the problem should be ignored.

For small sites, the impact will be marginal — Googlebot has plenty of time to crawl everything. For high-volume platforms, e-commerce, or media sites, it’s another story: each unnecessary recrawl is a missed opportunity to explore new pages or detect important updates.

  • Vary: User-Agent on a responsive site is an inconsistent configuration that generates superfluous recrawls
  • Indexing is not compromised, but crawl budget is wasted
  • High-volume sites are the most affected by this inefficiency
  • The fix is simple: remove the Vary header or configure it correctly according to the actual technical architecture
  • Google does not penalize this error, but it remains a hindrance to crawl optimization

SEO Expert opinion

Is this statement consistent with field observations?

Yes, completely. Technical audits regularly show sites serving the same responsive content with a Vary: User-Agent in place. In server logs, we see that Googlebot visits these URLs more frequently, alternating between mobile and desktop User-Agent, without impacting ranking or presence in the index.

What Mueller does not explicitly state is how significant this waste can be. On a site with 50,000 URLs and a limited crawl budget, a doubling of the necessary passes can delay the discovery of new content by several days. This is measurable in Search Console data and logs, but Google remains vague about the actual extent of the problem. [To be verified] according to your volume and update frequency.

What nuances should be added to this statement?

Mueller speaks of a responsive site, but the situation differs with dynamic serving (serving different HTML according to the User-Agent). In this case, the Vary: User-Agent header is not only recommended but mandatory for Google to understand that it needs to crawl both versions.

Thus, the problem does not arise from the Vary header itself, but from its use inconsistently with the technical architecture. If you are using responsive design, do not use Vary. If you are using dynamic serving, Vary is essential. Confusion often arises from poorly managed architectural changes: the site transitions to responsive but the HTTP headers remain configured for dynamic serving.

In what cases does this rule not apply?

If you are actually serving different content based on the User-Agent — for example, a mobile subdomain with a different HTML structure — then the Vary header is relevant and the recrawls are justified. Google needs to see both versions to index them correctly.

Another edge case: some CDNs or proxies automatically add Vary headers to manage their internal cache without the site owner being aware. In this case, the problem is not solved on the origin server but in the CDN configuration. An analysis of the actual HTTP responses sent by the CDN is necessary to identify the source of the problem.

Attention: Do not confuse the Vary header with the viewport meta tag or mobile annotation. These are three distinct mechanisms that do not serve the same functions. The Vary header relates to cache and crawl, not to device detection on the rendering side.

Practical impact and recommendations

How can you check if your site is sending an unnecessary Vary header?

Use your browser's DevTools (Network tab) or a tool like cURL to inspect the HTTP response headers of your main pages. If you see Vary: User-Agent while your site is responsive, you have identified the problem.

Check your server logs to see if Googlebot crawls your URLs alternating between mobile and desktop User-Agent over a short period. A systematic double crawl pattern is a good indicator that Google is reacting to this inconsistent Vary header. The Search Console will not directly alert you on this point — it's in the logs where this is visible.

What should you do concretely to correct this error?

If your site is truly responsive and serves the same HTML to all devices, remove the Vary: User-Agent header from your HTTP responses. On Apache servers, this is done in the .htaccess or vhost configuration. On Nginx, in the relevant server or location block.

If you are using a CDN (Cloudflare, Fastly, Akamai, etc.), check the cache configuration: some CDNs automatically add a Vary: User-Agent to manage their own caching policies. In this case, disable this rule or adjust the configuration so that it does not apply to crawlers. Then test with a crawler on your end to validate that the change has been properly implemented.

What errors should be avoided when making this correction?

Do not remove the Vary header if you are actually using dynamic serving — you would then break Google's ability to detect that multiple versions exist. Ensure first that your architecture is truly responsive before making changes to the headers.

Avoid modifying headers without testing the impact on the CDN cache. A sudden change may lead to a cache purge and a temporary increase in server load. Plan for a gradual rollout if you manage a high-traffic site. Finally, document the change in your technical procedures to avoid a future deployment inadvertently reintroducing the error.

  • Inspect HTTP response headers to detect an unnecessary Vary: User-Agent
  • Analyze server logs to identify patterns of double crawling by Googlebot
  • Remove the Vary header if the site is responsive (same HTML for all devices)
  • Check the CDN configuration to ensure it does not automatically add this header
  • Test the change with a crawler before deployment in production
  • Document the change to prevent future regressions
This optimization of HTTP headers requires advanced technical expertise, especially on complex infrastructures with CDNs and multiple environments. If you are not comfortable with this type of manipulation or if you seek support on all your crawl optimizations, engaging a specialized SEO agency can save you valuable time and secure your deployments.

❓ Frequently Asked Questions

L'en-tête Vary peut-il vraiment empêcher l'indexation de mes pages ?
Non. Mueller est clair : une mauvaise configuration de Vary ne nuit pas à l'indexation. Les pages restent indexées et classées normalement. Le seul impact est un gaspillage de crawl budget.
Comment savoir si Googlebot crawle mes pages en double à cause de Vary ?
Analysez vos logs serveur : si vous voyez des passages alternés avec User-Agent mobile et desktop sur les mêmes URLs à quelques heures d'intervalle, c'est un indicateur que Google réagit à un Vary: User-Agent.
Dois-je garder l'en-tête Vary si je fais du dynamic serving ?
Oui, absolument. Si vous servez un HTML différent selon le User-Agent (mobile vs desktop), l'en-tête Vary: User-Agent est indispensable pour que Google comprenne qu'il doit crawler les deux versions.
Mon CDN ajoute automatiquement un Vary: User-Agent, que faire ?
Vérifiez la configuration de cache de votre CDN et désactivez cette règle si votre site est responsive. Certains CDN permettent de configurer des règles spécifiques pour les crawlers.
Combien de temps faut-il pour que Google arrête les recrawls inutiles après correction ?
Cela dépend de la fréquence de crawl de votre site. Généralement, quelques jours à quelques semaines suffisent pour que Googlebot ajuste son comportement une fois l'en-tête Vary corrigé.
🏷 Related Topics
Crawl & Indexing AI & SEO Mobile SEO

🎥 From the same video 9

Other SEO insights extracted from this same Google Search Central video · duration 56 min · published on 04/09/2019

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.