BingYU wrote: > Thank God I find this message on the list > At the same time, I want push the question further. > > On Mon, Aug 15, 2005 at 07:57:58PM -0400, Roberto C. Sanchez wrote: > >>On Mon, Aug 15, 2005 at 07:42:24PM -0400, Hendrik Boom wrote: >> >>>A few weeks agol I think on this list, I saw a message asking >>>about software to lownload series of numbered files from a web site >>>-- series like >>> http://foo.bat/onion/soup01.htm >>> http://foo.bat/onion/soup02.htm > > let say, we have *not* like this soup01.. 02 03 > but we have onbaord.wmv someother.wmv > > I want to download the *.wmv files > http://www.via.com.tw/en/resources/download-center/videos/ > > I try to use > wget -c http://.../*.wmv > but it does not work >
Did you try the suggestion that you quoted further down in this message? (see my comment after that). > > > >>> http://foo.bat/onion/soup03.htm >>> http://foo.bat/onion/soup04.htm >>> >>>and the like, in situations where the site did *not* allow >>>you to read the directory >>> http://foo.bat/onion/ >>> >>>If it did allow such access, wget would obviously be the tool of >>>choice. >>> >>>I've tried looking in the archives for the message, but I can't find it. >>>I think the original poster wanted to download pictures, in case >>>it helps anyone remember. >>> >>>Anyway, I'd appreciate pointers to appropriate software >>>(or to the original message and responses). >>> >> >>I think you want curl. Likewise, this works: >> >>#!/bin/bash >>for i in `seq -w 1 10` ; do wget http://foo.bat/onion/soup$i.htm; done >> >>-Roberto >>-- >>Roberto C. Sanchez >>http://familiasanchez.net/~roberto Being that I know nothiung about scripting I rely on "DownThemAll!" ( http://downthemall.mozdev.org/) for such things. -- Scott www.angrykeyboarder.com © 2005 angrykeyboarder & Elmer Fudd. All Wights Wesewved