Understanding robots.txt
The robots.txt file sits at the root of your domain (e.g. example.com/robots.txt) and uses simple directives to communicate with bots. Here's what each directive means:
- User-agent — specifies which crawler the rules apply to.
*means all crawlers. - Disallow — tells crawlers not to access the specified path.
Disallow: /blocks the entire site. - Allow — explicitly permits a path, even if a parent directory is disallowed.
- Sitemap — points crawlers to your sitemap.xml for faster and more complete indexing.
