User-agent
Identifies which crawler a group of rules applies to.
Understand how the robots.txt file works, what it can and cannot do, and why it matters for search engines and AI crawlers.
No sign-up required. Runs directly in your browser.
Identifies which crawler a group of rules applies to.
Specify paths that bots may or may not crawl.
Optional pacing hint for how frequently bots should request pages.
robots.txt is a public text file that lives at the root of your domain, such as https://example.com/robots.txt. It tells crawlers which paths they are allowed to request and which ones they should avoid. It does not secure private data, but it acts as a first gate for crawlers discovering your site.
Identifies which crawler a group of rules applies to.
Specify paths that bots may or may not crawl.
Optional pacing hint for how frequently bots should request pages.
Points crawlers to one or more XML sitemaps for faster discovery.
Yes. A broad Disallow rule can prevent crawling. Testing helps prevent accidental lockouts.
No. robots.txt controls crawling, while noindex (meta robots) controls whether a page is indexed.
Many AI crawlers respect robots.txt as a public access policy, so it is a good place to define preferences.
Use the free Robots.txt Tester to fetch, parse and review crawl rules for your domain.
Test robots.txt with Robots.txt Tester