Hi, On Mon, 16 May 2011, Andrew Haley wrote:
> On 16/05/11 10:45, Richard Guenther wrote: > > On Fri, May 13, 2011 at 7:14 PM, Ian Lance Taylor <i...@google.com> wrote: > >> I noticed that buglist.cgi was taking quite a bit of CPU time. I looked > >> at some of the long running instances, and they were coming from > >> searchbots. I can't think of a good reason for this, so I have > >> committed this patch to the gcc.gnu.org robots.txt file to not let > >> searchbots search through lists of bugs. I plan to make a similar > >> change on the sourceware.org and cygwin.com sides. Please let me know > >> if this seems like a mistake. > >> > >> Does anybody have any experience with > >> http://code.google.com/p/bugzilla-sitemap/ ? That might be a slightly > >> better approach. > > > > Shouldn't we keep searchbots way from bugzilla completely? Searchbots > > can crawl the gcc-bugs mailinglist archives. > > I don't understand this. Surely it is super-useful for Google etc. to > be able to search gcc's Bugzilla. gcc-bugs provides exactly the same information, and doesn't have to regenerate the full web page for each access to a bug report. Ciao, Michael.