HTTrack Website Copier
Free software offline browser - FORUM
Subject: Re: Downloading picture files
Author: Stonn
Date: 04/04/2013 21:44
 
What if the article consists of dozens of pages? Copying them manually would
take a long time.
I hoped the program would find the links on the first page and download the
missing pages. (thus I added +*art52833* since every other page's link is made
up like /art52833-2.html then /art52833-3.html and so on)

Also, I still cannot get any images. Here is my command line:

(winhttrack -qwr1%e0C2%Ps2u1%s%uN0%I0p3DaK0H0%kf2A25000%f#f -F "Mozilla/4.5
(compatible; HTTrack 3.0x; Windows 98)" -%F "<!-- Mirrored from %s%s by
HTTrack Website Copier/3.x [XR&CO'2010], %s -->" -%l "en, en, *"
<http://pclab.pl/art52833.html> -O1 C:\Users\Michael\Desktop\PcLab\pad +*.html
+*.gif +*.jpg +*.png +*.tif +*.bmp )

Also, is it possible to put a variable into the scan rule? Such as
+/art52833-X.html would include every page no matter what X is. Or in the
beginning field with the links.

And, what does it mean to spider? I guess you mean scanning the page for links
but how do I ´´MUST let it spider the site to get them´´? Can I prevent
the program from spidering? I though this is what its made for - such that I
don't have to download every site separately.
 
Reply Create subthread


All articles

Subject Author Date
Downloading picture files

04/04/2013 03:50
Re: Downloading picture files

04/04/2013 04:17
Re: Downloading picture files

04/04/2013 21:44
Re: Downloading picture files

04/04/2013 22:24




6

Created with FORUM 2.0.11