robots.txt file tells search engines which pages can, or can't, be crawled on a site. It contains groups of rules for doing so, and each group has three main components:
- The user agent, which notes which crawler the group of rules applies to. For example,
- The rules themselves, which note specific URLs that crawlers can, or can't, access.
- An optional sitemap URL.
Shopify generates a default
robots.txt file that works for most stores. However, if you want to make changes to the default file, then you can add the robots.txt.liquid template in order to make the following customizations:
Add a new rule to an existing group
If you want to add a new rule to an existing group, you can adjust the Liquid for outputting the default rules to check for the associated group and include your rule.
For example, you can use the following to block all crawlers from accessing pages with the URL parameter
Remove a default rule from an existing group
If you want to remove a default rule from an existing group, you can adjust the Liquid for outputting the default rules to check for that rule and skip over it.
For example, you can use the following to remove the rule blocking crawlers from accessing the
Add custom rules
If you want to add a new rule that's not part of a default group, then you can manually enter the rule outside of the Liquid for outputting the default rules.
Common examples of these custom rules are:
Block certain crawlers
If a crawler isn't in the default rule set, you can manually add a rule to block it.
For example, the following would allow you to block the
Add extra sitemap URLs
The following example, where
[sitemap-url] is the sitemap URL, would allow you to include an extra sitemap URL: