more stats and analysis of robots.txt files were released in Google's Sitemaps
If your site has a robots.txt file, the new robots.txt tab provides Googlebot's
view of that file, including when the spider last accessed it, the status it
returns (404 if it does not exist), and if it blocks access to your home page.
It also shows any syntax errors in the file.
You can enter a list of URLs to see if the robots.txt file allows or blocks
You can also test changes to your robots.txt file by entering and then
testing them against the Googlebot user-agent, other Google user agents, or the
Robots Standard. This lets Google Sitemap users experiment with changes to see how they would
impact the crawl of your site, as well as make sure there are no errors in the
file, before making changes to the file on your site.
If you don't have a robots.txt file, you can use this page to test a potential
robots.txt file before you add it to your site.
Crawl stats now include the page on your site that had the highest PageRank, by
month, for the last three months.
Page analysis now includes a list of the most common words in your site's
content and in external links to your site. This gives you additional
information about why your site might come up for particular search queries.