Google published a new Robots.txt refresher explaining how Robots.txt enables publishers and SEOs to control search engine crawlers and other bots (that obey Robots.txt). The documentation includes examples of blocking specific pages (like shopping carts), restricting certain bots, and managing crawling behavior with simple rules.

From Basics To Advanced

The new documentation offers a quick introduction to what Robots.txt is and gradually progresses to increasingly advanced coverage of what publishers and SEOs can do with Robots.txt and how it benefits them.

The main point of the first part of the document is to introduce robots.txt as a stable web protocol with a 30 year history that’s widely supported by search engines and other crawlers.

Google Search Console will report a 404 error message if the Robots.txt is missing. It’s okay for that to happen but if it bugs you to see that in the GSC you can wait 30 days and the warning will drop off. An alterative is to create a blank…

Read Full Story: https://news.google.com/rss/articles/CBMijgFBVV95cUxQaUVKcFVvbFd3WHVxZnQxcEk4eTBzd0xhTjdZNlJUS3duRXBDRnFqdGgyM2ZJTGR5MVBoNmhxeDkyUDZFSW85SnZNcGhSMmtNcVAzR0s4ZFV0T3p0WjROSEJWNlhTeDNvbGszM0JLcmh5TjhzU0JDb1ZjenlhOVdKV2lpelp4ZVgxSlBvdGpn?oc=5