More stats and analysis of robots.txt files

February 06, 2006

Today, we released new features for Sitemaps.

robots.txt analysis

If the site has a robots.txt file, the new robots.txt tab provides Googlebot's view of that file, including when Googlebot last accessed it, the status it returns, and if it blocks access to your home page. This tab also lists any syntax errors in the file.

quick view of the analysed robots.txt file in webmaster tools

You can enter a list of URLs to see if the robots.txt file allows or blocks them.

analysis of multiple urls against the robots.txt file in webmaster tools

You can also test changes to your robots.txt file by entering them here and then testing them against the Googlebot user-agent, other Google user agents, or the Robots Standard. This lets you experiment with changes to see how they would impact the crawl of your site, as well as make sure there are no errors in the file, before making changes to the file on your site.

If you don't have a robots.txt file, you can use this page to test a potential robots.txt file before you add it to your site.

More stats

Crawl stats now include the page on your site that had the highest PageRank, by month, for the last three months.

report of the blocked urls with the highest pagerank in webmaster tools

Page analysis now includes a list of the most common words in your site's content and in external links to your site. This gives you additional information about why your site might come up for particular search queries.