| Hi,
I want to create a rip of the mail.google.com/support site, because it's
unreadable - I have to click 6 links to get to the actual information and then
go back and click again 6 times.
Anyway, I've written a simple *.bat script (on Windows) to concatenate a
webpage out of the *.html links provided. The *.html links though, must be in
the right order.
My problem is like this: I've analyzed the new.lst file and it appears, that
httrack downloads the link structure of a site like this:
--website
+-link1
+-link2
+-link3
...
+-linkm
+link1_1
+link1_2
+link1_3
+link1_4
...
+link1_m
and stores that structure in the new.lst file.
What I'd like to have is a fully 'recursive' structure, like this:
--website
+-link1
+link1_1
+link1_2
+link1_3
+link1_3_1
+link1_3_1_1
+link1_3_1_2
+link1_4
...
+link1_m
+-link2
+-link3
...
+-linkm
Is it somehow doable with httrack? When I wanted to download the google page,
I first thought that if I'll use maxConnectionsAtATime=1 the link structure
will be preserved. Furtunately for me, that setting does not interfere with
creating a consistent (see example 1) link structure (I think).
I've read the commandline manual and it appears that httrack has no option to
do what I want.
If that is the case, maybe someone knows a program that can help me to achieve
my task. Help is really appreciated.
Regards. | |