Ensure that search engines get the right indexing signals

Gain clarity on exactly which URLs are indexable and which are not, so you can make sure that only the right pages end up in the index.

Suganthan Mohanadasan

Suganthan Mohanadasan

Basic Gravity

With Sitebulb I was able to crawl large websites with thousands of pages effortlessly at a record time. It runs fast, and I love the javascript rendering options. The UI is very intuitive and gives a tonne of actionable hints, insights and recommendations. Apart from hints, I love the visual crawl map as it shows exactly how your pages are scattered. Finally, They have fantastic product support. If you’re an SEO person, you will fall in love with this tool.

Get to the bottom of any indexing issue

Sitebulb’s Indexability Report will help you untangle even the most complex indexing setups, giving you a clear understanding of anything that is going wrong. Whether it’s an over-zealous robots.txt file, conflicting noindex rules, or misplaced canonical tags, Sitebulb will alert you to any configuration issues.

See how robots directives impact indexability

Sitebulb breaks down indexing signals using language you can understand, splitting crawled URLs into two camps: Indexable and Not Indexable. Once you've identified an error or potential problem, you can dig in further to check noindex directives or canonical tag issues.

Robots directives affect on indexing

Robots.txt files can cause major issues in terms of crawling and indexing, as wayward disallow rules can prevent big chunks of a website from even being crawled. Sitebulb will tell you every single URL that is affected by the robots.txt file, and even pick out the specific robots.txt rule triggered.

Robots.txt issues

Quickly identify canonical tag issues

Sitebulb will allow you to quickly check any pages with canonical tags that are not self-referential, so you can double check these are pointing to the right URLs.

Further, it will check for canonical configuration issues, such as duplicate declarations or malformed URLs, and identify inconsistencies caused by compound robots rules – such as canonical tags that point at noindexed URLs.

Canonical issues

Avoid potential problems from duplicate robots declarations

Robots directives can be specified in 3 different locations – in the HTML <head>, the HTTP header, and on the robots.txt file. This can lead to multiple directives for a single URL, which potentially also leads to conflicting directives. These type of inconsistencies are typically very hard to identify manually, and can cause massive problems if left unchecked.

Sitebulb will pull all these rules together and automatically check them all, picking out the specific issue and all URLs affected.

Ready to try Sitebulb?
Start your free 14 day trial now

Start Free Trial