HTTrack Website Copier
Free software offline browser - FORUM
Subject: Re: No search for duplicate links
Author: WHRoeder
Date: 05/05/2013 04:03
 
1) Always post the ACTUAL command line used (or log file line two) so we know
what the site is, what ALL your settings are, etc.
2) Always post the URLs you're not getting and from what URL it is
referenced.
3) Always post anything USEFUL from the log file.
4) If you want everything use the near flag (get non-html files related) not
filters.
5) I always run with A) No External Pages so I know where the mirror ends.
With B) browser ID=msie 6 pulldown as some sites don't like a HTT one. With C)
Attempt to detect all links (for JS/CSS.) With D) Timeout=60, retry=9 to avoid
temporary network interruptions from deleting files.

> As the topic says, can I set WinHTTrack to not
> search for duplicate links? 
Since HTT doesn't search for anything your question makes no sense. The
default is to download all links it finds (recursively) in the starting URL.

> This page is similar to a wiki page, and some words
"This page?" What page? There are no mind readers here.

> is a link to its article, and many words appear on
> many articles. So can I set Winht to not search for
> duplicate links? This would speed up the process
> increasingly.
It doesn't search! And if it finds a duplicate URL, it's only going to
download it once. There's nothing to speed it up.

> The site in question is <http://www.newadvent.org/> (I
> would have all the pages from this domain, but you
> can get a better example from
> <http://www.newadvent.org/cathen/05677a.htm> )
Make up your mind which URL you're downloading!
 
Reply Create subthread


All articles

Subject Author Date
No search for duplicate links

05/05/2013 00:41
Re: No search for duplicate links

05/05/2013 04:03




6

Created with FORUM 2.0.11