HTTrack Website Copier
Free software offline browser - FORUM
Subject: Re: crawl single folder but save all external images
Author: WHRoeder
Date: 12/23/2012 18:22
 
1) Always post the ACTUAL command line used (or log file line two) so we know
what the site is, what ALL your settings are, etc.
2) Always post the URLs you're not getting and from what URL.
3) Always post anything USEFUL from the log file.
4) If you want everything use the near flag (get non-html files related) not
filters.
5) I always run with No External Pages so I know where the mirror ends. I
always run with broswer ID=msie6 as some sites don't like a HTT one. I always
run with Attempt to detect all links.

> other domains. How do I limit only to this folder
> and sub-folders, and how do I mirror images (but

Nothing. The default is to stay on site and go down only.

> ONLY images, nothing else) that are linked to other
> domains? In my last experiment, I left HTTRACK

Can't be done. If you don't have the URLs of the images, you MUST let it
spider the site to get them: -* +*.html +*.jpg

> running overnight, and it was crawling entire
> websites because my HTML files have URL links to
> other websites, but all I really need mirrored are

Then you did something that allowed it. No mind readers here, #1
 
Reply Create subthread


All articles

Subject Author Date
crawl single folder but save all external images

12/23/2012 17:50
Re: crawl single folder but save all external images

12/23/2012 18:22
Re: crawl single folder but save all external images

12/23/2012 19:42
Re: crawl single folder but save all external images

12/23/2012 22:35
Re: crawl single folder but save all external images

12/23/2012 23:07
Re: crawl single folder but save all external images

12/24/2012 03:16




9

Created with FORUM 2.0.11