| 1) Always post the ACTUAL command line used (or log file line two) so we know
what the site is, what ALL your settings are, etc.
2) Always post the URLs you're not getting and from what URL it is
referenced.
3) Always post anything USEFUL from the log file.
4) If you want everything use the near flag (get non-html files related) not
filters.
5) I always run with A) No External Pages so I know where the mirror ends.
With B) browser ID=msie 6 pulldown as some sites don't like a HTT one. With C)
Attempt to detect all links (for JS/CSS.) With D) Timeout=60, retry=9 to avoid
temporary network interruptions from deleting files.
> i've need absolutely a copy offline of a website but
> impossible copy link on this page (*.htm?ca=2_s)
That's not a page. #2
> <http://www.leboncoin.fr/ventes_immobilieres/offres/a>
> quitaine/dordogne/?f=a&th=1&pe=2&q=maison&it=1
The default is down only and
/ventes_immobilieres/463135685.htm?ca=2_s
is not down from
/ventes_immobilieres/offres/aquitaine/dordogne
Either enable bidirection or enable all ventes_immobilieres
+*/ventes_immobilieres/*
> I need this page and page result (advertisement +
> pictures + description and phone number)
> I've trying with including +*[name].htm?ca=2_s/* but
> nothing.
Since the default is to download everything, that enabling filter does
nothing.
What did the log file say about robots.txt and what did you do about it. | |