Google has published one of those technical notes that looks niche at first glance but is actually quite useful for ordinary businesses with ordinary websites. In a new Search Central blog post, Google explained that Googlebot usually fetches up to 2MB of an individual HTML page before it stops. For most businesses in Bath, Somerset, Wiltshire, Dorset, Bristol and Gloucestershire, that is not a reason to panic. But it is a useful reminder that bloated pages, messy templates and overstuffed site builds can quietly get in the way of search visibility.
The key point is simple. If the important bits of your page sit too far down in the raw HTML because the page is padded with huge blobs of code, inline images, heavy scripts or sprawling builder markup, Google may not fetch all of it. And if Google does not fetch it, it cannot render or index it properly. That matters most for SEO-led pages, service pages, local landing pages and any content meant to bring in enquiries.
What Google has actually said
In the official post, Google said Googlebot is not really a single neat little robot any more. It sits on top of a wider crawling system used by different Google products. For Google Search specifically, the important figure is that Googlebot currently fetches up to 2MB for a normal URL, excluding PDFs, which have a much larger limit. Google’s crawler overview documentation also notes that the wider infrastructure has a default 15MB file-size limit, but individual crawlers can set their own smaller thresholds.
A page can still load normally for a human while being less tidy from Google’s point of view. Once the fetch hits the cutoff, the bytes beyond that point are simply not processed.
Why this matters for local businesses rather than just SEO nerds
Plenty of South West businesses now run websites built from page builders, layered plugins, tracking tools, cookie systems, chat widgets, booking tools and design extras that have accumulated over time. None of that is automatically bad. The problem starts when the invisible weight at the top of the page gets so chunky that the useful parts arrive late in the source.
A Bath solicitor, a Bristol retailer, a Dorset trades firm or a Somerset charity might all have the same underlying issue: the site looks polished, but the structure under the bonnet is doing them no favours. Important headings, service explanations, canonicals, schema markup or local trust signals can end up buried lower than they should be. If you already care about SEO in Bath or across the wider South West, this is exactly the sort of quiet technical housekeeping that is worth paying attention to.
It is also a reminder that redesigns are not always improvements. A shiny new theme or a pile of extra scripts can add far more code than most businesses realise.
Who should worry, and who probably should not
Google itself says that 2MB of HTML is massive for most pages, and many smaller brochure sites will never come close. So this is not an emergency for everyone.
But it is more relevant if your site has any of the following:
- heavy page-builder layouts with lots of nested blocks and modules
- large amounts of inline CSS or JavaScript
- base64-encoded images or decorative assets embedded directly into the page
- very large menus, filter systems or repeated template elements before the main content
- service or ecommerce pages that rely on scripts to reveal important text later
Those are common enough that the update is still useful for local firms, especially ones investing in new content, local SEO pages or lead-generation landing pages. Good website content writing and editing helps, but the technical delivery matters too. A clear page still needs to be fetchable in a sensible way.
What to check next
First, check whether your important pages are unusually bloated.
Your developer can inspect the raw HTML size and see whether the page source is far heavier than it needs to be. This is about the actual delivered code, not just how pretty the page looks once everything has loaded.
Second, make sure critical information appears early enough in the source.
Titles, meta information, canonicals, structured data, headings and the start of the main content should not be pushed miles down by avoidable clutter.
Third, move weight out of the HTML where sensible.
Google’s advice was clear here: heavy CSS and JavaScript are usually better kept in external files, because each fetched resource has its own limit. That often gives your main HTML more room to carry the genuinely important signals.
Fourth, be wary of gimmicky page bloat.
Huge decorative sections, giant hidden panels, duplicated modules and overengineered templates can all make a page harder to process without making it more useful for visitors.
The practical takeaway
The calm takeaway is that most businesses do not need to lose sleep over a 2MB Googlebot cutoff. But if your site has grown messy, script-heavy or builder-heavy, Google has given a clear explanation of why that can matter. For local organisations around Bath and the surrounding counties, it is worth remembering the next time a redesign, SEO tidy-up or content refresh is on the list.
Sources:
Google Search Central Blog — Inside Googlebot: demystifying crawling, fetching, and the bytes we process
Google for Developers — Overview of Google crawlers and fetchers

