| > > All I want is that downloading <http://www.google.com>
> should
> > get me:
> >
> > index.html
> > images/logo.gif
> >
> > That's it. But no combination of flags I use in httrack
> (on
> > Windows or Linux!) seems to get me that.
> Well, I just tried getting www.google.com with WinHTtrack
> 3.23 (beta), let it go for about a minute, and it easily
> crawled 40+ pages with images, including that logo.gif
> mentioned above. However I noticed these problems (some
Well yes, that's my problem too. Either too much (way too
much, depending on the page) or too little (no images).
The in-between "sweet spot" is missing :-)
Oh well... I guess I have to give up. Mind you, I love
lwp-rget, but the parallelisation here was lovely! Too bad...
Thanks anyway,
Sitaram
PS: If anyone is interested, here's why I need this
particular type of mirroring. I do a lot of surfing
"offline" using a 128MB USB flashdrive which I hook up
wherever I go. If I come across a URL I want, I copy it
into a text file on the same flashdrive. About once a week
I'll use the fast internet at work and download all the URLs
copied so farin one shot (typically, in a week of sporadic
offline browsing, about 50-80 URLs). Yes, this means my web
browsing is very "jerky", if you see what I mean, but I dont
mind! Saves a lot of time for me to do it this way. | |