What does Google say about SEO? /
Quick SEO Quiz

Test your SEO knowledge in 3 questions

Less than 30 seconds. Find out how much you really know about Google search.

🕒 ~30s 🎯 3 questions 📚 SEO Google

Official statement

Google's crawl infrastructure has a default 15 megabyte size limit. When this limit is reached, the crawler stops receiving data. This limit is set at the infrastructure level and applies to all crawlers that do not override it.
🎥 Source video

Extracted from a Google Search Central video

💬 EN 📅 12/03/2026 ✂ 12 statements
Watch on YouTube →
Other statements from this video 11
  1. Googlebot est-il vraiment un seul programme ou une infrastructure distribuée ?
  2. Le crawl Google fonctionne-t-il vraiment par API avec des paramètres configurables ?
  3. Pourquoi Google ne documente-t-il pas tous ses crawlers dans sa liste officielle ?
  4. Crawlers vs Fetchers : pourquoi Google utilise-t-il deux systèmes distincts pour accéder à vos pages ?
  5. Google réutilise-t-il vraiment le cache entre ses différents crawlers ?
  6. Pourquoi Googlebot crawle-t-il principalement depuis les États-Unis ?
  7. Pourquoi Google ne crawle-t-il pas massivement votre contenu géobloqué ?
  8. Pourquoi le géoblocage peut-il nuire au crawl de votre site par Google ?
  9. Le crawl budget est-il vraiment protégé automatiquement par Google ?
  10. Pourquoi Google impose-t-il une limite de 2 Mo pour crawler vos pages web ?
  11. Pourquoi Google limite-t-il le crawl des PDFs à 64 Mo alors que le HTML plafonne à 2 Mo ?
📅
Official statement from (1 month ago)
TL;DR

Google enforces a default 15 MB limit on all crawled pages. Beyond that threshold, the crawler stops receiving data. This limit is applied at the infrastructure level to all Google crawlers, except where explicitly overridden.

What you need to understand

What does this 15 MB limit actually mean in practice?

When Googlebot requests a page, it begins receiving HTML, CSS, JavaScript, or other content. If the total file size exceeds 15 megabytes, the crawler cuts the connection and stops receiving the remaining data.

This limit is coded at the crawl infrastructure level, not at the algorithmic level. It therefore applies to all Google bots (Googlebot Desktop, Mobile, Ads, News...) unless a specific crawler overrides it — which happens rarely and opaquely.

Why does Google enforce this constraint?

Two main reasons: resource economy and abuse prevention. Crawling billions of pages is expensive in bandwidth and time. Limiting file sizes prevents a single domain from monopolizing resources by serving massive files.

Second point — a 15 MB page is already enormous for standard HTML. If your page reaches this size, you have a structural problem: bloated JavaScript, base64-encoded inline images, redundant CSS.

Does this limit affect indexation?

Yes, directly. If Googlebot stops receiving data at 15 MB, everything after that point is invisible to the index. Internal links, text content, structured markup — everything that appears after the cutoff is never seen.

For dynamic sites that load content via JavaScript, it's even more problematic. If the main JavaScript exceeds 15 MB (poorly optimized bundles, bloated vendors), the rendering engine cannot execute it completely.

  • Universal limit: 15 MB applies to all Google crawlers by default
  • Hard cutoff: no negotiation, the crawler stops receiving data abruptly
  • Indexation impact: all content beyond 15 MB is lost for Google
  • Infrastructure level: technical decision, not algorithmic
  • Override possible: certain crawlers can exceed this limit, but it's rare and undocumented

SEO Expert opinion

Is this 15 MB limit consistent with practices observed in the field?

Yes and no. Historically, Google has always had file size limits — this isn't new. Official documentation has mentioned a « several megabytes » limit for years without ever giving a precise figure.

What's new is the explicit confirmation: 15 MB. Before, we were flying blind. Field tests suggested a limit between 10 and 20 MB, but nothing official. Gary Illyes finally puts a number on the table.

Now, the real question — how many sites actually exceed 15 MB per page? Honestly, very few. The median webpage in 2025 hovers around 2-3 MB (HTTP Archive). Sites exceeding 15 MB are generally pathological cases: e-commerce with thousands of inline products, poorly optimized SPAs, unpaginated documentation pages.

What gray areas remain in this statement?

First point — Gary says « unless a crawler overrides it ». [Needs verification] Which crawlers? Under what circumstances? Google provides no details.

We can assume that certain specialized crawlers (Googlebot-Image, Googlebot-Video) have different limits tailored to their use case. But this is an assumption, not documented fact.

Second gray area — does this limit apply only to the initial HTML or also to loaded resources (external CSS, JS)? The phrasing « the crawler stops receiving data » suggests it applies at the level of each individual HTTP request. If your HTML page is 2 MB but your JS bundle is 16 MB, that bundle will be truncated.

Warning: Sites using heavy JavaScript frameworks (Next.js, Angular with large vendors) must absolutely verify their bundle sizes. A 15+ MB JS file can compromise Google's rendering.

Does this limit pose a real problem for most sites?

No, let's be honest. If your page is 15 MB, you already have a performance problem well before worrying about crawling. Users on mobile or slow connections will never see your content.

That said, certain sectors are more at risk: news portals with dozens of widgets, technical documentation sites with inline code, SaaS applications with complex dashboards. For these cases, the 15 MB limit can become an invisible ceiling.

Practical impact and recommendations

How do I check if my site exceeds the 15 MB limit?

First step: audit the size of your pages and resources. Use Chrome DevTools (Network tab) to measure the total size of each HTTP request. Filter by type (Document, Script, Stylesheet) and check individual files.

To automate auditing at scale, tools like Screaming Frog or Sitebulb let you crawl your site and extract the size of each resource. Export the data and sort by descending size — you'll immediately spot outliers.

Attention — don't limit yourself to HTML. Also check your JavaScript bundles and CSS stylesheets. A vendor.js file of 18 MB will be truncated, potentially breaking rendering on Googlebot's side.

What optimizations should I apply if resources exceed 15 MB?

First lever: code splitting. If you use a modern framework (React, Vue, Angular), divide your JavaScript bundles into smaller chunks loaded on demand. Webpack, Vite, or Rollup do this natively.

Second action — eliminate dead code. Tree shaking, removing unused vendors, lazy loading non-critical components. An audit with Lighthouse or Bundle Analyzer often reveals dozens of imported but never-used libraries.

For inline images (base64), stop immediately. Images must be served as separate files, optimized to WebP or AVIF, with lazy loading. A single 4K base64-encoded image can blow up your HTML size.

If you have listing pages with hundreds of products, implement strict pagination or infinite scroll with progressive loading. Google crawls 10 pages of 50 products better than a single page of 500 products.

Should I monitor this metric continuously?

Yes, especially if your site evolves rapidly. Add automated monitoring of resource size into your CI/CD pipeline. Define alert thresholds: if a bundle exceeds 10 MB, block deployment until investigation.

On the SEO side, integrate this check into your quarterly audits. Dynamic sites with multiple dev teams can see their weight explode without oversight — a plugin added here, a third-party lib there, and suddenly your page is 20 MB.

  • Audit the size of all strategic pages with DevTools or Screaming Frog
  • Check JavaScript and CSS bundle sizes — alert threshold at 10 MB
  • Implement code splitting to divide large JS files into smaller chunks
  • Remove unused dependencies via tree shaking and bundle analysis
  • Convert all inline images (base64) to external optimized files
  • Paginate or lazy-load product lists or voluminous content
  • Set up automatic alerts in CI/CD if a resource exceeds 10 MB
  • Review rendering architecture for SPAs: SSR or pre-rendering if needed
The 15 MB limit is not a problem for most well-optimized sites. But for complex platforms with heavy JavaScript or dynamically generated content, it's a ceiling worth monitoring closely. If your technical architecture makes these optimizations complex to implement, partnering with a specialized SEO agency can help you identify bottlenecks and deploy solutions tailored to your technology stack.

❓ Frequently Asked Questions

La limite de 15 Mo s'applique-t-elle au HTML seul ou à toutes les ressources ?
La limite s'applique à chaque requête HTTP individuelle. Si votre HTML fait 2 Mo mais qu'un fichier JavaScript externe fait 16 Mo, ce fichier JS sera tronqué à 15 Mo. Chaque ressource est limitée indépendamment.
Que se passe-t-il exactement quand Googlebot atteint la limite de 15 Mo ?
Le crawler arrête immédiatement de recevoir les données. Tout contenu au-delà de ce seuil est invisible pour l'indexation — liens internes, balises meta, texte, scripts. C'est une coupure brutale, pas une négociation.
Peut-on demander à Google d'augmenter cette limite pour un site spécifique ?
Non, cette limite est codée au niveau infrastructure et s'applique universellement. Google ne propose aucun mécanisme officiel pour demander une exception. La seule solution est d'optimiser la taille de vos ressources.
Les images comptent-elles dans les 15 Mo si elles sont chargées via des balises <img> ?
Non, les images externes chargées via <img> ou CSS background-image ne comptent pas dans la limite de la page HTML elle-même. Seule la taille du fichier HTML (et les ressources inline comme base64) est concernée.
Comment savoir si Google a tronqué mes pages à cause de cette limite ?
Aucun signal direct dans Search Console. La seule façon est de comparer la taille de vos ressources (DevTools Network) avec ce que Google affiche dans le cache HTML ou l'outil d'inspection d'URL. Si du contenu manque, c'est probablement une troncature.
🏷 Related Topics
Crawl & Indexing Pagination & Structure

🎥 From the same video 11

Other SEO insights extracted from this same Google Search Central video · published on 12/03/2026

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.