HTTrack Website Copier
Free software offline browser - FORUM
Subject: Get only near links in ouside domains.
Author: Mark
Date: 09/21/2015 18:41
 
I've been trying to get a website that has its indexes on the same level as the
content it links to, but in a different folder. To use the example from the
FAQ:

I'm downloading:
www.someweb.com/gallery/trees
It stores a series of html-files that contain descriptions of trees as well as
links to pictures of those trees. These pictures are located in:
www.someweb.com/photos

Now, if I add
+www.someweb.com/photos/*
to my Scan Rules, it'll get everything in that folder, including the pictures
of flowers, which I don't need. I just want the pictures that are linked to
from /trees.

If I allow 1 level of outside links, it'll also get the items from
www.someweb.com/advice
which contains advice on how to trim those trees, which I don't want. Adding
-*
will exclude these, but then forces the use of
+www.someweb.com/photos/*
with the above issue.

So I want only the content of /trees and whatever files in /photos these link
to, no other photos.

I figured I would have it by now, but even Google doesn't seem to know, and
I'm pretty good with search terms.
 
Reply


All articles

Subject Author Date
Get only near links in ouside domains.

09/21/2015 18:41
Re: Get only near links in ouside domains.

09/21/2015 21:25




e

Created with FORUM 2.0.11