Hello I want to download a large amount of html files from a web server that supports http gzip compression. I tried to call wget with the --headers option to enable the compression:
wget -nc -np -k -r --header\="Accept-Encoding: gzip" http://address The compression is enabled, and the first downloaded document is stored unter the file name it has on the server. However, it is still compressed with gzip, which prevents wget from looking up links in that document for recursive downloading. Now I am looking for a way to either tell wget to compress the files when writing them to disk, or to find the links in the compressed files. Of course I could uncompress the first file, restart wget, uncompress the next set of files and so on, but I would like to use a way that is more convenient. Any ideas? best regards Andreas Janssen -- Andreas Janssen [EMAIL PROTECTED] PGP-Key-ID: 0xDC801674 Registered Linux User #267976 -- To UNSUBSCRIBE, email to [EMAIL PROTECTED] with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]