Closed pimoore closed 3 months ago
I did some looking into this and the only answers I got were:
The order of directives in a robots.txt file, including the placement of the Sitemap directive, generally does not affect the interpretation by compliant web crawlers.
The directive containing the sitemap location can be placed anywhere in the robots.txt file. It is independent of the user-agent line, so it does not matter where it is placed.
I suspect that robots.txt is working as expected, but I do agree that it's certainly more clear when it's not in an explicit disallow list for User-Agent: *
.
I've just pushed an update.
Lighthouse testing is showing the robots.txt file is malformed and not being downloaded, which would result in it not being applied at all. Here's the current format:
I believe the Sitemap field should be last, as the convention expects the Disallow variables to immediately follow the User-Agent.