Google recently updated its technical documentation to clarify the file limits crawled by Googlebot. The limit is now set to 2 MB for each file type supported by Google Search (HTML, CSS, JavaScript, etc.), compared with 15 MB mentioned previously in the old documentation. PDF files, for their part, retain a limit of 64 MB.
However, as several Anglo-Saxon SEO experts have pointed out, this is not a sudden policy change but rather a documentation clarification that now distinguishes the limits of Google's different crawlers. The 15 MB limit remains the default limit for Google's general crawling infrastructure, while Googlebot applies a stricter 2 MB limit when crawling for indexing in Google Search.
A documentation clarification, not a new change
According to Search Engine Journal, this update is part of a broader reorganization of Google's documentation that began in late 2025. Barry Schwartz of Search Engine Roundtable notes that these limits may have existed for a long time, but are now officially documented more clearly.
John Mueller, at Google, confirmed on Reddit that " " 2 MB of HTML is already a lot " " and that it is extremely rare for sites to encounter problems with this limit. According to HTTP Archive data cited by Mueller, about 90% of web pages have less than 151 KB of HTML.
What actually changes
This 2 MB limit applies to each file crawled individually (and not to the page as a whole), notably:
- HTML files
- CSS stylesheets
- JavaScript files
- Other formats supported by Google Search
PDF files benefit from an exception with a limit kept at 64 MB. Important to note: this restriction applies to the uncompressed data, and each resource referenced in the HTML (CSS, JS) is fetched separately with its own 2 MB limit.
When Googlebot reaches this 2 MB limit, it stops the download and only indexes the portion of the file that has already been retrieved. The rest of the content is simply ignored.
According to the official Google documentation :
« When crawling for Google Search, Googlebot fetches the first 2 MB of a supported file type and the first 64 MB of a PDF file. From a rendering point of view, each resource referenced in the HTML code (CSS or JavaScript, for example) is fetched separately, and each resource fetch is subject to the same file size limit that applies to other files (with the exception of PDF files). Once the limit is reached, Googlebot stops fetching and only submits the portion already downloaded of the file for indexing. »
Real impact: most sites are not affected
In practice, is it that serious? No, for the overwhelming majority of websitesAccording to the HTTP Archive Web Almanac, the median size of an HTML file is about 33 KB based on the data cited by John Mueller, or roughly 60 times smaller than the 2 MB limit. Additionally, 90% of web pages have less than 151 KB of HTML.
To put things in perspective: 2 MB of plain text equals about 2 million characters, the equivalent of several hundred pages of a standard novel. You really have to go out of your way to create a page that large.
Which sites are affected?
Some sites using slightly too heavy CSS or JavaScript frameworks can, however, be impacted. The sites most likely to exceed 2 MB are those that:
- Integrate large amounts of JavaScript directly into the HTML (rather than via external files)
- Use large, unoptimized CSS libraries
- Bundle many dependencies into a single file
- Embed large JSON objects inline for e-commerce sites
- Use single-page applications (SPAs) that load entire JavaScript frameworks inline
What to do if your site is affected?
If you manage a site with potentially heavy files, here are the recommendations:
- Check the size of your files : Use Chrome DevTools (Network tab), Screaming Frog SEO Spider, or Google Search Console to identify pages exceeding 1.5 MB
- Externalize scripts and styles : Move inline CSS and JavaScript to external files, so each resource benefits from its own 2 MB limit
- Split large pages : Split long content into logical sections with a good internal linking structure
- Minify your code : Remove whitespace, comments and unnecessary characters to reduce file sizes
- Implement lazy loading : Defer loading of content at the bottom of the page
- Monitor indexing : Use Search Console to detect possible visibility changes
A performance optimization approach
This documentation clarification is part of a web performance optimization approach, particularly important for mobile users where overly large files can harm the browsing experience, especially on limited connections.
Moreover, this limit aligns with best practices for Generative Engine Optimization (GEO) : most crawlers artificial intelligence (ChatGPT, Perplexity) do not render JavaScript and consume only raw HTML. A structured, lightweight, semantic HTML therefore improves visibility both in traditional search and in AI-powered engines.
Concerned? Request a full audit of your site!
- 1Website
- 2Number of URLs
- 3Your requirement
- 4Your specifics
- 5Your contact details
- 6Estimate
Abondance's expertise is:
+2,000 audits Delivered for SMEs, mid-market companies and large enterprises.
15+ years of experience Senior consultants & Abondance ambassadors.
Proven method Prioritized roadmap, concrete actions.The article "Google clarifies Googlebot's crawling limits: 2 MB per file" was published on the site Abondance.