Crawlers on Gemini and best practices

Stephane Bortzmeyer stephane at sources.org
Thu Dec 10 14:15:54 GMT 2020


On Thu, Dec 10, 2020 at 03:07:50PM +0100,
 Côme Chilliet <come at chilliet.eu> wrote 
 a message of 4 lines which said:

> I don't see anything in the spec saying to stop at first match. I think you should read the whole response and apply all lines that matches your virtual user agent.
> So in this case for an archiver, all lines.

Then this example in <http://www.robotstxt.org/robotstxt.html> would
not work:

User-agent: Google
Disallow:
User-agent: *
Disallow: /

Because with your algorithm, Google would be disallowed (while the
comment in the page says "To allow a single robot[Google]").



More information about the Gemini mailing list