Package: moinmoin-common Version: 1.5.3-1 Severity: wishlist Tags: patch I have had an annoyance with one of my MoinMoin-based wikis, and that is that I cannot use wget to retrieve pages in raw or printable format.
Turns out this is an apparently deliberate measure taken on MoinMoin/request.py, to keep robots from doing harmful things -- however, it appears to be the case that the list of not-harmful things that a robot can do has increased without the code reflecting this. I have attached a patch that I've tested. It works fine for me. -- System Information: Debian Release: testing/unstable APT prefers unstable APT policy: (500, 'unstable'), (500, 'testing'), (500, 'stable') Architecture: powerpc (ppc) Shell: /bin/sh linked to /bin/bash Kernel: Linux 2.6.14-2-powerpc-smp Locale: LANG=C, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
--- MoinMoin/request.py 2006-04-15 12:21:49.000000000 -0400 +++ MoinMoin/request.py.new 2006-07-06 19:12:14.000000000 -0400 @@ -883,9 +883,12 @@ qs = self.query_string if ((qs != '' or self.request_method != 'GET') and not 'action=rss_rc' in qs and - # allow spiders to get attachments and do 'show' + # Allow spiders to get attachments view the page in any format. not ('action=AttachFile' in qs and 'do=get' in qs) and - not 'action=show' in qs + not 'action=show' in qs and + not 'action=raw' in qs and + not 'action=print' in qs and + not 'action=format' in qs ): forbidden = self.isSpiderAgent