HTTrack Website Copier
Free software offline browser - FORUM
Subject: Re: recursive scanning is a problem
Author: Filer
Date: 12/16/2002 22:13
 
I have no idea really, try setting the Travel Mode to can go down only. If you
know how, you might want to check the responses the web server gives for
requesting the files, what is the date it gives for the file? (telnet to the
http port of the server and issue the GET command manually, I'm afraid I
cannot give more precise help than this)

A misconfigured web server might give the revision time for the file as
current time, whether this would be a misconfiguration or done on purpose hard
to say.

This would cause Httrack to download the file always again because it would
seem that the file has changed. If there is no way to limit the download based
on filters or depth, you might be out of luck.

Sometimes I find that in order to download a certain site I have to enter
multiple subdirectories into the URL box instead of the server root, because
of the way the site is built - much more easy than writing separate scan rules
for all possible instances.

Very often 'when there's a will there's a way', except with the sites using
URLs obfuscated by complex javascript functions. Wish someone will tackle
these some day.
 
Reply Create subthread


All articles

Subject Author Date
recursive scanning is a problem

11/29/2002 01:25
Re: recursive scanning is a problem

11/29/2002 07:37
Re: recursive scanning is a problem

12/13/2002 06:55
Re: recursive scanning is a problem

12/16/2002 22:13
Re: recursive scanning is a problem

12/23/2002 18:14
Re: recursive scanning is a problem

02/16/2003 15:04
Re: recursive scanning is a problem

03/01/2003 22:24




2

Created with FORUM 2.0.11