April 9, 2024

What is Crawl Management?

Once you own a website, or start a job working behind the scenes building or optimising websites, you understand that they work a little differently than you might have once thought. 

Rather than just existing on the web and showing up on Google when searched for, websites must first be ‘crawled’ and indexed by search engines to show up on their results at all. ‘Crawling’ happens frequently over the lifespan of a website to help search engines such as Google understand the content on the site, and add to and refresh their indexed results. These crawlers, also known as ‘bots’ or ‘spiders’ also crawl your website to view it once it has been updated, recently become popular, or if it is frequently linked to elsewhere through backlinks. This can be requested through Google Search Console, or regularly as the site becomes discovered by Google. 

Crawling a site from the bots’ perspective is time and energy-consuming. However, they will continue to crawl the site regularly if they believe that there is a good enough chance that will find something new on the site or page. 

Effective SEO might mean that you need to manage these bots’ time and energy better, ensuring that your site gets added to Google’s index more efficiently than just allowing the bot to run its course unguided. This is called ‘crawl management’. So what is ‘crawl management’ exactly, and how can you use it to get the best SEO from your campaign?

‘Crawl Management’ is a method used by SEOs and web developers to limit the number of pages that are effectively crawled and indexed by Search Engines to only relevant URLs. This may mean excluding certain pages or parts of your site from being crawled or improving your site structure and technical background to allow for more efficient crawling by bots such as Googlebot. 

Find out everything you need to know about website crawling and how to efficiently request crawling and indexing from the team at Wildcat Digital. 

Jump to:

How Does Crawling Work?

First off, the internet is huge.

Google, in its seemingly infinite capacity, attempts to understand the internet better and provide better results constantly. However, as it is essentially a service in itself, it only tries to understand the part of the internet that is going to be usable and beneficial to its users. To do this, it crawls the internet to find, read, understand and index sites and pages that its users may want to find.

As an SEO agency, it’s our job to help Google and other Search Engines find sites and pages to add to this index. Then, using SEO, tools, techniques and clever methods, we get Search Engines to promote our sites further up in their Search Engine Results Pages (SERPs). These are rankings, and they are how Search Engines promote sites within their index that they believe offer the best service to their users. 

These indexed pages are around 30 – 50 billion in number, with trillions existing elsewhere on the net. It has been estimated that around 4% of the internet is actually indexed by Google, and therefore eligible to be found in Google Search Results. This is how it can find so many results in such a short period of time. 

Crawling is therefore the process of discovering page content, including copy, images, documents and links to other web pages, which are followed to discover new URLs (unless robots are otherwise instructed). Crawling the web is a task that Google takes very seriously, and it does so within an allocated budget. 

What is a Crawl Budget?

A ‘crawl budget’ is the number of pages that Google’s crawling bot (Googlebot) is willing to visit on a specific website within a specific time period. According to Google’s own documentation, this is defined by two main elements: crawl capacity limit and crawl demand.

As Googlebot can’t see all of the internet all of the time, it must allocate its budget of resources to websites that are worth serving to its users. Think of a crawl budget as an allowance for sites that meet Google’s requirements, or have asked for indexing through Google’s Search Console

The crawl budget can be influenced by a variety of factors including the website’s quality as seen by the search engine, server performance, update frequency, page quality, relevance and how frequently the website is updated compared to your other sites. At its essence, more crawl budget is allocated out to websites that might require it, over ones that don’t. If you are allocated more crawl budget, more of your web pages will be crawled more frequently. 

Looking at the factors above, you can make improvements to the website to improve the crawl budget. This might be why you might want to employ an SEO agency to work for you. Certainly, the work that SEOs do would help bots crawl your site more often and crawl more pages when it does so. 

What Factors Influence Crawl Budget? 

Googlebot and other crawlers are programmed to crawl your website without causing any strain on your servers. In order to achieve this, Googlebot determines a crawl capacity limit, which is the maximum number of parallel connections that can be used to crawl a website at the same time, along Google’s crawling limits.

If you have a large site that contains hundreds or even thousands of pages, then having it crawled regularly may slow down your site. If it is badly linked internally then it may cause a strain on the crawlers when attempting to crawl. If your site is old and not updated regularly or doesn’t get viewed often this may lead to problems convincing the crawlers to prioritise it over fresher content elsewhere. 

A variety of factors influence crawl budget, some of which we have mentioned above, but others might include: 

Server Performance

Slow servers or server response times can greatly hinder crawling efficiency. Additionally, if your pages repeatedly return 5xx errors (server-side error), search crawlers are likely to drop these results from rankings and even slow down the crawling of the website. If you own a website, or you are an SEO looking to test your server response time, you can use tools such as GTMetrix. Tools like Google Search console and website crawlers like Screaming From and Sitebulb can be used to find and analyse 5xx errors on your website.

Robots.txt

Directives in the robots.txt file guide crawlers to sitemaps, and allowed and disallowed pages. Setting up a robots.txt page can give instructions to bots crawling your site about where to go or not go next. This can help you manage your crawl budget efficiently. 

XML Sitemaps

XML Sitemaps should exist in their own right but also be linked to your robot.txt page to instruct bots on which pages exist on your site. XML Sitemaps act as a directory for users and bots to quickly find all relevant URLs associated with a site and crawl accordingly.

These help search engines discover pages efficiently. You can submit a sitemap to Google Search Console which will be read every few days and allow Googlebot to discover new pages on your website.   Find out how often you should update your sitemap (and how…) 

Duplicate Content

Pages with duplicate content are a waste of your crawl budget. And, when indexed, they may also create cannibalisation issues by competing with each other. Duplicate content may occur due to issues with pagination, filtering, spammy page duplication, product variations or instances where content is available on the site and through a download (such as a printout or PDFs).

What’s more, duplicate pages without proper robots markup, may lead to Googlebot drawing its own conclusions about canonicals, and excluding certain pages from showing up on search results. More on how to address duplication issues below.

4xx & Soft 404 Errors

Broken links waste crawl budget. Once a crawler gets to a link it wants to crawl the content, find new links and continue doing its job. If it gets to a broken link, like a 4xx then it ends its journey. 

Once Google encounters a page with a 404 Permanently removed code, it will stop coming back to crawl that page. However, when crawlers encounter soft 404 pages, they continue crawling these, potentially wasting the crawl budget.  

Having broken links on your website then not only wastes a good opportunity for your users but also presents a risk to having your website effectively crawled, too. 

Low-Value Pages

Low-value pages present an issue to effective crawling as they essentially waste the crawl budget. Over time, if the content on the site is generally poor, irrelevant, or outdated, Google may reduce the crawl budget on the site.  

Redirect Chains

When a page is redirected, it will return a 301 code to the crawler, pointing it to a new URL and therefore issuing a new crawl request for the same page. Redirect may build over time, creating redirect chains, meaning that the crawler has to hop from URL to URL multiple times before reaching the destination URL. 

This is not only a waste of crawl resources, but it also increases the page’s load time, and as we have mentioned above, slow site performance may lead to a loss of crawl budget over time. What’s more, Googlebot only follows 5 redirect hops before giving up on the crawling of the page, meaning that the final URL may never be crawled and indexed. 

How Do You Manage a Crawl Budget?

Using the above as a guide, we can improve our allocated crawl budgets by improving various aspects of the site. 

Improve Server Performance and Page Load Times

Improving server response and page load times can be carried out by an SEO, web developer or webmaster. Tools like WP Rocket for WordPress can help defer loading speeds, compress images, minify CSS and Javascript or introduce lazy loading can help bots and users use your site more efficiently. If your site responds quickly to Google’s requests, the allocated crawl capacity limit goes up and more of your site can be crawled. 

Disallow Irrelevant pages in robots.txt

Disallowing the craw of irrelevant pages, page paths, directories, or parameters found on your site can vastly reduce the number of pages that are findable to crawlers and free up the crawl budget for more important URLs. Do so carefully as some pages may internally link to other important pages that you want Google to index. Find more about how robots.txt works here

Update your XML Sitemap

Your Sitemap ought to automatically generate through Yoast or a similar tool. Most CMSs will do this for you. 

If your site already has an XML sitemap, then ensure that it is updated regularly. Make sure to review your settings to ensure that only relevant URLs are included in your sitemap. 

Remove Duplicated Content

To avoid wasting your allocated crawl budget, it is important to manage duplicate pages with appropriate canonicalization. Paginated URLs, filtered dynamic URLs, product variations, and duplicated pages, among others, should be canonicalised to the main URL, in order to consolidate your content and ensure that only relevant pages are crawled and indexed. 

Address 404 Internal Links and Remove soft 404s

Marking permanently removed pages with a 404 may be the best bet for some, but as many pages are internally linked and may even be ranking, setting up relevant redirects for pages that have been removed will allow your crawl to continue and potentially find new areas of your site yet to be crawled. 

As we have mentioned above, soft 404s waste your crawl budget. These are usually either pages with errors that still return a 200 OK response or pages with no content. You should monitor soft 404s in Google Search Console and make sure to implement 404 or 410 responses for permanently removed pages, redirect to the correct resource where relevant, or populate your page with the correct content and ensure that it returns and 200 OK response

Remove or Optimise Low-Value Pages

Ensuring that your website is populated with relevant and valuable content and regularly updated, is a good way of sending search engines the positive signal they need to recognise the value of your pages and allocate more crawl budget to your site over time.

Google Search Console may provide some insight into the pages that are deemed low-value in the ‘Crawled – currently not indexed’ section of the ‘Pages’ page. 

If these pages don’t currently serve a purpose and are not within the inner workings of your site, you could remove them to improve crawlability elsewhere. 

Remove Redirect Chains

Redirects have their place in any SEO strategy, but redirect chains should be avoided for all of the reasons mentioned above. To remove redirect chains, update your redirect to point to the final URL. You should also ensure that the URLs within the chain are not linked to internally on your website. 

Improving Your Website’s Crawlability With Wildcat Digital

If your website is struggling with crawling and indexing new URLs, you may have an issue with its crawlability. Getting your site crawled and indexed can be tricky, especially if you are already running a business yourself. 

In most cases employing an SEO through an agency is a cost-effective way to get your website seen by Google and billions of Google users, helping you compete in the digital world. 

Wildcat Digital is an SEO, PPC and Paid Social Digital Marketing Agency based in Sheffield, Yorkshire. We specialise in helping businesses punch above their weight online and compete with some of the largest names in their fields. 

If you need help with crawling and indexing your website, then get in contact with our team today

Post by

Jon Herdman

More blogs.

View all