Configuration
Tell crawlers where your sitemap is. Highly recommended.
How long crawlers should wait between requests. Most sites don't need this.
Paths you want to block from crawlers (e.g. /admin/, /private/, /api/)
robots.txt

What Is a Robots.txt File?

A robots.txt file tells search engine crawlers which pages or sections of your website they can and cannot access. It lives at the root of your domain (e.g., https://example.com/robots.txt) and is one of the first files crawlers check before indexing your site.

Common Robots.txt Rules

Blocking AI Crawlers

Many website owners now choose to block AI training crawlers from scraping their content. Common AI crawler user agents include GPTBot (OpenAI), CCBot (Common Crawl), and Google-Extended (Google AI training). Use the checkbox above to add these rules automatically.

Best Practices