Package: www.debian.org
User: www.debian....@packages.debian.org
Usertags: packages
Severity: important

Hi,

While looking at blocking some web spiders that were causing
significant load on picconi (packages.d.o master), I noticed that the
robots.txt currently doesn't attempt to block them.

Specifically, the current file is:

<quote>
User-agent: *
Disallow:
</quote>

which is effectively the same as allowing everything. "Disallow: /"
might be more logical, unless there is a desire / requirement to allow
crawling and indexing of (parts of) the site.


Regards,

Adam

Reply via email to