I have a project where I wish to download all of a core website, plus the
single destination page of each outgoing link on the site.
The doit file is:
-qw%e1C2%Ps2u1Z%s%uN0I0%I0p1BaK4H0%kf2A25000%f#f -F "Mozilla/4.5 (compatible;
HTTrack 3.0x; Windows 98)" -%F "<!-- Mirrored from %s%s by HTTrack Website
Copier/3.x [XR&CO'2008], %s -->" -%l "en, en, *" www.discoverireland.ie -O1
"C:\\My Web Sites\\www.discoverireland.ie"
and I would expect that httrack would not scan any of the external pages.
However, I am finding it does scan each external page, and continues to
download further linked page, so that I get lines like this in the new.txt
file, where <http://www.failteireland.ie> is one of the external files with link
So you can see it appears to be extending beyond the link depth of one.
Am I misunderstanding what the external link depth actually means?