Doing SEO for ecommerce? Our upcoming webinar is just for you! Register now.

How to crawl large websites

The aim of this post is to provide a methodology for crawling large websites with Sitebulb.

Crawling large websites is a tricky subject, primarily because of the number of unknowns. Until you actually crawl a website, you don't know if you're working with a 1,000 page website or a 100,000 page website. And that is before you start thinking about embedded resources, external links and subdomains.

But when you are dealing with a big website, crawl data can be incredibly valuable, as it can reveal patterns and issues that you'd struggle to detect otherwise.

So that leads us on to an important question:

How large is a large website?

This can depend on perspective. An experienced enterprise SEO who is familiar with 1 million+ page websites might see a 5000 page site as tiny. But to a solo in-house SEO in their first job, it can feel enormous.

For the sake of argument, we'll draw the line at 100,000 URLs. This is the point at which you might need to start thinking about which crawl and analysis options you have switched on in Sitebulb.

In general, for sites smaller than 100,000 URLs, you should be pretty safe turning on whichever crawl options you like.

Worked example: Patient

We'll work through an example, say I needed to crawl the Patient website, which is a UK based health advice site, for doctors and patients alike.

If I were actually working with the client, one of my initial Q & A questions would be to ask them about scale, but since I'm not, we'll lean on our friend Google.

Google results

First things first, this does NOT mean that we will need to crawl exactly 302,000 pages. Not included in this total are noindexed URLs, disallowed URLs, canonicalized URLs, page resource URLs, external links or links to subdomains - yet all of this stuff could end up in the scope of your Audit if you are not careful.

If anything, what Google displays can only really be considered a lower bound, for the purposes of planning a crawl.

At this point we know it's a pretty big site, so our first port of call is a Sample Audit.

Running a Sample Audit

The point of a Sample Audit is to crawl a small subset of a website, in order to get a feel for how a full Audit will go.

We first need to set up a new Project in Sitebulb. You will be able to select 'Sample' as your Audit Type at this point. Once selected, the sample crawl settings will appear, which is how we will limit the crawl.

In this case, we are only going to crawl 10 levels deep, and a maximum of only 1500 URLs at each level (Sitebulb will choose 1500 random URLs to crawl at each level).

Setting up a Sample Audit

Once we hit 'Save and Continue', Sitebulb will go off and perform a number of 'pre-audit checks', such as checking the robots.txt file to make sure we can actually crawl the website in the first place. 

The default setting is to use the HTML crawler and collect Page Resources, so we will not change any of this, then set the audit running.

Crawl Progress

If you keep an eye on the speed, this will give you an idea how fast the site can be crawled, and therefore how long it might take to complete the main audit. In this case we're at around 8 URLs/second, which is relatively fast. You can also experiment with ratcheting the speed up and seeing how fast it will comfortably go (watch out for errors creeping in, this typically means you are going too fast).

Viewing the Sample Audit

Once the Sample Audit has finished running, we can use the data collected to make inferences about how a full Audit would work on the site.

Sample audit results

One thing we can clearly see is that there is roughly one page resource URL crawled, for every internal URL crawled.

This means that if we included Page Resources in our main Audit we would roughly double the amount of URLs we need to crawl, so 300,000 suddenly turns into 600,000.

There was also roughly one external URL crawled for every internal one, so across 300,000 internal pages this would add another 300,000 external URLs.

The point in doing this is to help us build up a profile of the website crawl to answer the question: 'what would it look like to do a full crawl of everything on the site?'

Then we will get an idea of how long it will take, to help decide if we are willing to wait this long for all the data.

Another element worth considering is how many indexable pages our sample crawl turned up, which we can get from the Indexability report:

Indexability data

The data in this report is potentially game changing - this is because all our assumptions are based off how many URLs are currently indexable. So, for example, if you came across 2 'Not Indexable' pages for every one that is indexable, your crawl would likely be three times bigger than expected.

In this case, it's not such a massive problem, but we still have one Not Indexable page for every 10 Indexable.

We can now build up our profile of what a full Audit might look like:

  1. We have a baseline of around 300,000 URLs from Google site: search
  2. If we decide to crawl page resources, this will add another 300,000 URLs (1 to 1)
  3. External URLs add another 300,000 URLs (1 to 1)
  4. Not Indexable URLs would add another 30,000 URLs (1 to 10)

So if we wanted to crawl everything, we'd be looking at crawling around 930,000 URLs. At a rate of 8 URLs/second, this would take about 32 hours to complete.

Bear in mind that this is all based on using the HTML Crawler. If you want to use the Chrome Crawler, it will take a LOT longer, and you would need to repeat the above exercise using your Chrome results to reach a useful conclusion.

Limiting the Audit

At this point we have a fairly good estimate for how long a full Audit would take, with the HTML Crawler. 

If we wanted to move forward and carry out the full Audit, we could ask ourselves if we really want to wait this long, or if we'd be comfortable omitting some of the data.

There's no way to not crawl 'Not Indexable URLs' (because Sitebulb only knows they're not indexable once it has crawled them...), but we can exclude external URLs and page resources, which would keep the total down closer to the 300,000 we started with.

To do this, we'd start a new Project and this time select 'Standard Audit' from the dropdown instead of 'Sample Audit.'

Excluding Page Resources

You could stop Sitebulb from crawling page resources altogether by disabled this setting. Alternatively, open the Advanced Settings panel for more granular control of the type of resources you want Sitebulb to crawl. 

Page Resources Advanced Settings

*Note: the Performance & Mobile Friendly report requires Page Resources to be enabled.

Excluding External Links

You can disable External Link analysis in the Search Engine Optimisation Advanced Settings, to stop Sitebulb form Checking the HTTP status of links to external websites. 

Disabling External Link Analysis in Search Engine Optimisation Advanced Settings

 You can also stop Sitebulb from crawling subdomain URLs by unticking this option in the Subdomain Options tab. 

Excluding Subdomain URLs

Disabling both External URLs and Subdomain checks is the best way you can limit the size of the Audit, without deliberately excluding internal pages in the site. Of course, whether you are willing to do so will depend on your website configuration and how valuable this data may be if included. 

Excluding Selected Internal HTML URLs

Finally, you can also exclude internal URL paths and/or parameterized URLs through the Advanced Settings. We've written a separate guide which covers that topic, as it can be quite involved: Limiting the Crawler for Faster and Cleaner Audits.

Pause and Resume

This final note is nothing to do with how you arrange the crawl setup, but more to do with how to manage your computer. Ideally, you want to leave Sitebulb to run continuously, complete the crawl and generate all the reports - this is where Sitebulb Cloud is the ideal choice. 

However, with bigger Audits ran on Sitebulb Desktop it may not be possible to leave your computer on for a long period (e.g. needing to shut down your machine overnight), so the best option is to utilise the 'Pause' feature, to make sure any crawl interruptions are controlled.

While an Audit is in progress, just hit the purple Pause button in the top right.

Pause a running audit

Once you've done this, wait around 5 seconds for the purple button at the top left to change from 'Pausing' to 'Paused.'

At this point you'll notice that there is now an option to 'Resume' in the top right.

Resume a paused audit

Once an Audit has been paused, you can close Sitebulb down and shut your machine down. Then when you reopen Sitebulb the next day, you'll see a message informing you that you have an incomplete audit, so you can jump back into it and get it going again.

To see a list of paused audits, click to view the Paused tab:

Paused Audits

Hit 'View Progress' to return to the progress screen, where you can then hit 'Resume' to set the Audit running again.

If pausing is not an option, and/or you regularly need to crawl large sites continuously for several days, Sitebulb Cloud may be a better choice for you an your team. It allows you to run large audits uninterrupted and at scale, as well as collaborate on shared projects. More about Sitebulb Cloud here

A Word on Disk Space

A final consideration when crawling large websites on Desktop is to make sure you have enough space on your hard drive for the audit. One of the reasons Sitebulb can crawl so many pages is because it writes the data to disk instead of holding it in RAM.

A site with 100,000+ URLs could take up anything from 250 MB to around 2 GB. There's no easy way to know how much space you'll need before you start the Audit, but bear in mind that the more data you collect, the more disk space will be required.

In particular, 'Link Analysis' can take up a lot of disk space, as the numbers can grow huge very quickly.

To put this into perspective, I crawled a site recently with 1.6 million internal URLs. I crawled it once with 'Link Analysis' switched off, and this took up 6 GB of space on my hard drive. I crawled it again and switched on 'Link Analysis', and it was 36 GB!

Why so large? It had 142,600,000 links! That's why.


  • Do a sample Audit.
  • Analyse the sample Audit to understand potential timeframes and limitations.
  • Adjust your crawl settings to limit the Audit (optional)
  • When running a full audit, pause and resume if you are unable to leave the crawler on continuously.
  • Make sure you've got enough disk space on your machine.