HTTrack Website Copier
Free software offline browser - FORUM
Subject: Wikis, images, and -n / --near
Author: Cédric
Date: 12/09/2019 21:28
 
Hello,

I'm trying to save a Wikipedia article, with just a depth of 1 for now (might
be more later), but I'd like to save its embedded images as well, and I can't
seem to understand how to do that, or why it doesn't work with the options I
use.

Let's say I want to save the page <https://fr.wikipedia.org/wiki/Etterbeek>,
with a command line like "httrack <https://fr.wikipedia.org/wiki/Etterbeek> -O
mirror --depth=1 --near". The page is saved, but the images are missing. When
I look at the original page's source code, I see that the images are stored on
<https://upload.wikimedia.org>. So it's a different domain. But with other
sites, as soon as I use the --near option, such external resources are
downloaded as well, even if I don't add any filters or options likes
--go-everywhere.

For example, on another wiki, when I try to mirror the page
<https://fr.wikimini.org/wiki/Herg>é, if I don't use the --near option, the
images are missing (they're stored on another subdomain). But as soon as I add
that option, the images are downloaded.

Has anyone any idea why it does not work on Wikipedia ?
Thanks.
 
Reply


All articles

Subject Author Date
Wikis, images, and -n / --near

12/09/2019 21:28




d

Created with FORUM 2.0.11