This may or may not help. There are programs that will capture a website. There is also software that will grab the content from a webpage and also grab content from any pages that have a link on the original webpage. Generally you can set the software to go X number of links deep off of each page. No idea how long it took to write something like that. From your log files, you could possibly use such a piece of software to grab the content of pages listed in the logs.
A recent post gave the setting to retrieve cached hits, that someone wanted to use for disaster recovery. This will only let you get the paged cached though, not every webpage someone requested. (not all pages get cached). Chris Perreault -----Original Message----- From: Harry Prasetyo K [mailto:[EMAIL PROTECTED] Sent: Monday, June 07, 2004 5:23 AM To: [EMAIL PROTECTED] Subject: RE: [squid-users] Reading Cache of my Squid Well, forgive me for my poor knowledge. Therefore i missused words for explaining my problems. Well how about fetching the URL objects that have been cached in the squid cache ? How squid can manage the request that is need to forward to internet or to look into the cache ? Or anyone knows some program (source code) on C or Perl which can do it ? i'll be thankful for any help ... HP > > Squid doesn't know what a webpage is; only browsers have an idea > about that. > Keep that in mind, squid caches url objects (only). > Pehaps this viewpoint implies changing your requirement(s) or whatever > your needs are. > > M. > > -- ** Architecture And Network Laboratory **
