HTTrack Website Copier
Free software offline browser - FORUM
Subject: Bug? Failure to only get HTML files with filters
Author: Harry Aep
Date: 04/02/2010 21:58
 
I know about the priority option to only get HTML files.
As a test,
I am attempting to only get HTML files with filters.
HTTrack only processes 1 file with these rules.

I place this .httrackrc in the output folder:
clean
robots 0
deny *
allow html
allow mime:text/html
allow mime:application/xml

Should it not retrieve all HTML files?robots.txt processing is disabled.
I tested on multiple web sites.
httrack is started as in
httrack www.httrack.com

from
hts-log.txt 
HTTrack3.43-5+libhtsjava.so.2
HTTrack Website Copier/3.43-5 mirror complete in 1 seconds : 1 links scanned,
1 files written ... bytes transfered using HTTP compression in 1 files, ratio
37%


The single HTTP page is retrieved successfully.
 
Reply


All articles

Subject Author Date
Bug? Failure to only get HTML files with filters

04/02/2010 21:58
Re: Bug? Failure to only get HTML files with filters

04/02/2010 22:27
Re: Bug? Failure to only get HTML files with filters

04/02/2010 22:48
Re: Bug? Failure to only get HTML files with filters

04/02/2010 22:52
Re: Bug? Failure to only get HTML files with filters

04/03/2010 10:19
Re: Bug? Failure to only get HTML files with filters

04/03/2010 10:20




2

Created with FORUM 2.0.11