What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 5 questions

Less than a minute. Find out how much you really know about Google search.

🕒 ~1 min 🎯 5 questions

Official statement

Server speed affects the number of pages Google can crawl each day. If a server is slow, Google may crawl it less frequently to prevent overloading.
11:48
🎥 Source video

Extracted from a Google Search Central video

⏱ 57:34 💬 EN 📅 18/10/2018 ✂ 9 statements
Watch on YouTube (11:48) →
Other statements from this video 8
  1. 8:11 Où placer vos données structurées pour qu'elles comptent vraiment ?
  2. 10:25 Google indexe-t-il vraiment toutes les pages qu'il explore ?
  3. 22:16 Les canonicals sont-elles vraiment évaluées comme les balises noindex par Google ?
  4. 23:49 Le JavaScript bloque-t-il vraiment l'indexation de vos pages par Google ?
  5. 31:39 Faut-il regrouper vos petits sites en un seul domaine pour améliorer votre SEO ?
  6. 34:39 Le Dynamic Rendering est-il encore une solution viable pour gérer le JavaScript en SEO ?
  7. 42:00 Faut-il vraiment optimiser toutes vos images pour Google Images ?
  8. 52:11 Faut-il vraiment corriger toutes les erreurs 404 dans Search Console ?
📅
Official statement from (7 years ago)
TL;DR

Google adjusts its crawling frequency based on your server's response speed. A high response time leads Googlebot to space out its visits to prevent overloading the infrastructure. Essentially, a slow server mechanically reduces the number of pages crawled daily, which particularly penalizes sites with high volumes or frequently updated content.

What you need to understand

Why does Google slow down crawling on a slow server?

Googlebot applies a logic of crawl rate limiting to protect the servers it explores. When a site takes several seconds to respond, Google interprets this as a sign of limited infrastructure capacity. The bot then automatically reduces its pace to avoid causing overload or timeouts.

This approach aims to preserve both the user experience of the crawled site and the efficiency of Googlebot. Fast crawling on a slow server poses risks of 503 errors, blank pages, and lost legitimate requests. Google prefers to space out visits rather than degrade site availability.

What is the actual impact on the number of crawled pages?

The crawl budget is not a fixed value: it fluctuates based on the server's capacity and the content's appeal. A server responding in 200 ms can handle 10,000 pages crawled per day, while a server with a 2-second response time caps at 1,000 pages for an equivalent level of solicitation.

High-volume sites (e-commerce, directories, media with deep archives) face a mechanical bottleneck. If your catalog contains 50,000 products and Google only crawls 800 pages per day, your new product listings take weeks to be indexed. The issue is not the content's priority but the available bandwidth.

How does Google detect a server's speed?

Googlebot measures the time to first byte (TTFB) for each request and aggregates this data over several days. It's not a one-time test but a moving average. An occasional slow spike is not enough to trigger a lasting reduction in crawling, but a high median over 7 days certainly is.

5xx errors also play a role: too many 503 or 500 errors indicate an unstable server. Google decreases the crawl rate even if response times remain acceptable, based on precautionary principles. Server logs and Search Console can help trace these adjustments.

  • Crawl rate limiting: Google automatically reduces its pace if the server shows signs of weakness (high TTFB, 5xx errors)
  • Direct proportionality: a server 10 times slower = crawl budget divided by approximately 10, all else being equal
  • Continuous measurement: median TTFB over several days, not a single snapshot
  • Maximal impact on large sites: a catalog of 100,000 pages with a slow server may only have 2-3% of its content crawled daily
  • No manual exceptions: even an authoritative site experiences this limitation if its infrastructure is underpowered

SEO Expert opinion

Does this statement align with real-world observations?

Yes, and it’s a rarely contested point within the SEO community. Technical audits consistently confirm the correlation between TTFB and crawl volume. An e-commerce client reduced their TTFB from 1.8 seconds to 400 ms and saw their daily crawl triple in 10 days without any other changes. Apache/Nginx logs tell the story.

The important nuance: this limitation primarily affects high-volume sites. A blog with 200 articles can have a mediocre server without visible consequences, as Google crawls the entire site in just a few minutes even at a reduced pace. The issue becomes critical when the total pages / daily crawl ratio exceeds 30 days.

What gray areas remain in this statement?

John Mueller provides no figures, no thresholds. [To be verified] What TTFB triggers a reduction in crawl? 500 ms? 1 second? 2 seconds? It's impossible to know officially. Empirical observations suggest that above a median of 800 ms, Google starts to throttle back, but this varies depending on the type of site and its overall authority.

Another unclear point: how does Google balance server speed against content importance? Does a slow site with breaking news or viral events receive temporary leniency? Nothing in official statements confirms this, but some real-world cases suggest that a manual override remains possible for extraordinary situations.

Should TTFB or overall infrastructure be prioritized?

TTFB remains the primary marker, but it's only a symptom. A server might respond quickly on the homepage but collapse on complex product listings with 12 SQL queries. Google crawls the entire site, not just the fast pages. Therefore, optimization should cover all types of pages, not just strategic templates.

Underpowered infrastructure leads to erratic TTFB. A CDN improves static distribution but doesn’t fix a backend server that’s too slow to generate HTML. The real question is: can your server handle 20 simultaneous requests from Googlebot without exceeding 500 ms? If not, the infrastructure is the bottleneck, not the technical stack.

Practical impact and recommendations

How can you diagnose a server speed issue affecting crawling?

First reflex: Google Search Console > Crawl Statistics. Look at the "Page Download Time (ms)" chart over 90 days. If the median exceeds 800 ms or shows an upward trend, you have a problem. Cross-reference with the volume of pages crawled: a simultaneous drop in crawl + rise in TTFB confirms the diagnosis.

Second lever: raw server logs. Extract Googlebot user-agents and calculate the average TTFB by page type (categories, product listings, articles). Use AWStats, GoAccess, or a custom Python script. Third-party tools like Oncrawl or Botify automate this analysis if you have the budget. The goal is to identify the sections of the site that hinder crawling.

What optimizations yield the quickest gains?

Start with the application server: upgrade from PHP-FPM 4 workers to 16, or migrate to a 4 vCPU instance if you're on a 1 vCPU VPS. Gains are immediate and measurable. One client went from 1.2 seconds to 350 ms TTFB just by doubling the RAM allocated to MySQL. Underpowered infrastructure is the number one cause.

Then move to application caching: Varnish, Redis, or even a simple static HTML cache via a WordPress plugin. Google often crawls the same URLs (pagination, product sorting): serving from cache reduces load by 60-80%. Ensure you properly configure Cache-Control headers so that Googlebot respects the cache without indexing outdated content.

What common mistakes exacerbate the problem?

A classic mistake: misconfigured CDN that speeds up the front-end but does nothing to TTFB for HTML. Cloudflare or AWS CloudFront don’t change anything if your origin server takes 2 seconds to generate the page. The CDN serves assets (JS, CSS, images) quickly, but Googlebot still waits 2 seconds for the HTML. A false impression of speed.

Another trap: optimizing only the strategic pages visible in GA. Google also crawls archives, old categories, and deep pagination pages. If 80% of your site has a catastrophic TTFB but your homepage and your 10 top products are fast, the overall crawl remains throttled. Improvement must be systemic, not cosmetic.

  • Measure median TTFB via Search Console and server logs across the entire site, not just popular pages
  • Size infrastructure to handle 20-30 simultaneous Googlebot requests without exceeding 500 ms
  • Implement application caching (Varnish, Redis, static HTML cache) on the most crawled page types
  • Monitor 5xx errors: more than 2-3% server errors trigger a preventive reduction in crawl
  • Test server-side HTML generation time with tools like Blackfire.io or New Relic APM
  • Verify that the CDN is not masking a TTFB server issue: test directly against the origin IP
A slow server mechanically reduces crawl budget, directly impacting the speed of indexing new pages. Gains can be measured in days: server optimization = doubling of daily crawl in 7-10 days. Prioritize infrastructure and application caching over CDN and front-end optimizations. These technical interventions often require specialized expertise in infrastructure and monitoring. If your internal team lacks resources or DevOps skills, enlisting an SEO agency specialized in technical performance can significantly accelerate diagnosis and the implementation of sustainable solutions.

❓ Frequently Asked Questions

Un CDN comme Cloudflare améliore-t-il le crawl budget ?
Non, pas directement. Le CDN accélère la distribution des assets statiques (images, CSS, JS) mais ne réduit pas le temps de génération HTML côté serveur, qui est le principal facteur mesuré par Google. Si votre TTFB reste élevé sur le HTML, le crawl budget reste limité même avec un CDN.
À partir de quel TTFB Google réduit-il la fréquence de crawl ?
Google ne communique aucun seuil officiel. Les observations terrain suggèrent qu'au-delà de 800 ms de TTFB médian, une réduction progressive du crawl s'opère. Au-delà de 1,5-2 secondes, la limitation devient très marquée, surtout sur les sites à forte volumétrie.
Un site d'autorité élevée bénéficie-t-il d'une tolérance sur la vitesse serveur ?
Partiellement. Un site d'actualité majeur ou une référence sectorielle conserve un crawl plus fréquent qu'un site lambda à TTFB équivalent, mais la limitation liée à la vitesse serveur s'applique quand même. L'autorité n'annule pas la contrainte technique, elle l'atténue.
Les erreurs 5xx ont-elles le même impact qu'un serveur lent ?
Oui, voire pire. Un taux d'erreurs 5xx supérieur à 2-3 % déclenche une réduction préventive du crawl, même si les pages qui répondent le font rapidement. Google interprète les erreurs serveur comme un signal d'instabilité et réduit la charge par précaution.
Peut-on demander manuellement à Google d'augmenter le crawl rate ?
Non, Google a supprimé cette option dans la Search Console. Le crawl rate s'ajuste automatiquement en fonction de la vitesse serveur, de la fraîcheur du contenu et de l'autorité du site. Seule l'optimisation technique côté serveur permet d'augmenter durablement le volume de crawl.
🏷 Related Topics
Domain Age & History Crawl & Indexing AI & SEO Web Performance

🎥 From the same video 8

Other SEO insights extracted from this same Google Search Central video · duration 57 min · published on 18/10/2018

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.