In the previous posts about the Robots Exclusion Protocol (REP) we explored what's already possible to do with its various components — namely robots.txt and the URI level controls. In this post we will explore how the REP can play a supporting role in the ever-evolving relation between automatic clients and the human web.
Category Archives: Google Webmaster Central Blog
Get Ready for Search Central Live Asia Pacific 2025
Hello 2025! (Yeah, we know, time flies!) We've had some exciting plans in the works for Search Central Live (SCL) Asia Pacific this year, and we're super excited to let you in on what we've been up to. We've been listening closely to your feedback, and we're cooking up something different than what we usually do–something bigger, deeper, and more tailored to you!
Source: Google Search Central Blog
Robots Refresher: page-level granularity
With the robots.txt file, site owners
have a simple way to control which parts of a website are accessible by crawlers.
To help site owners further express how search engines and web
crawlers can use their pages, the web standards group came
up with robots meta
tags in 1996, just a few months after meta
tags
were proposed for HTML (and anecdotally, also before Google
was founded). Later, X-Robots-Tag
HTTP response headers were added.
These instructions are sent together with a URL, so crawlers can only take them into account
if they're not disallowed from crawling the URL through the robots.txt file. Together, they
form the Robots Exclusion Protocol (REP).
Source: Google Search Central Blog
Robots Refresher: robots.txt — a flexible way to control how machines explore your website
A long-standing tool for website owners, robots.txt has been in active use for over 30 years and is broadly supported by crawler operators (such as tools for site owners, services, and search engines). In this edition of the robots refresher series, we'll take a closer look at robots.txt as a flexible way to tell robots what you want them to do (or not do) on your website.
Source: Google Search Central Blog
Robots Refresher: robots.txt – a flexible way to control how machines explore your website
A long-standing tool for website owners, robots.txt has been in active use for over 30 years and is broadly supported by crawler operators (such as tools for site owners, services, and search engines). In this edition of the robots refresher series, we'll take a closer look at robots.txt as a flexible way to tell robots what you want them to do (or not do) on your website.
Source: Google Search Central Blog
Search Central Live is going to Madrid
We're very excited to announce that Search Central Live is going to Madrid for the first time on April 9! The event will have a mix of presenters from the Google Search, News, and Partnerships teams and the content will be delivered in English and Spanish, but we'll have live translation.
Source: Google Search Central Blog
First Search Central Live in South Africa
On April 2, 2025 we'll be in Johannesburg, South Africa for the very first Search Central Live event in Africa! We're excited to welcome you at Search Central Live South Africa and talk about all things Google Search!
Source: Google Search Central Blog
Robots Refresher: introducing a new series
Every now and then we get questions about robots.txt, robots meta tags, and the control functionality that they offer. Following our December series on crawling, we thought this would be the perfect time to put together a light refresher. So, if you're curious about these controls, follow along in this new blog post series!