| I try to capture the following website:
www.forum.jadekwan.net
There is no error reported in the statistics while capturing. Here is the
log:
HTTrack3.33+swf launched on Sat, 08 Oct 2005 17:59:08 at
<http://www.forum.jadekwan.net> +*.png +*.gif +*.jpg +*.css +*.js
-ad.doubleclick.net/*
(winhttrack
-qir20%e5C1%Pns2u1%s%uN0%I0p3DaK0c32T1200R6H0%kf2A25000#L100000%f0#f -F
"Mozilla/4.5 (compatible; HTTrack 3.0x; Windows 98)" -%F -%l "en, en, *"
<http://www.forum.jadekwan.net> -O "C:\Jade Forum Site\Jade,C:\Jade Forum
Site\Jade" +*.png +*.gif +*.jpg +*.css +*.js -ad.doubleclick.net/* -%A
php3,php,php2,asp,jsp,pl,cfm,nsf=text/html )
Information, Warnings and Errors reported for this mirror:
note: the hts-log.txt file, and hts-cache folder, may contain sensitive
information,
such as username/password authentication for websites mirrored in this
project
do not share these files/folders if you want these information to remain
private
17:59:11 Info: Note: due to www.forum.jadekwan.net remote robots.txt rules,
links begining with these path will be forbidden: /admin, /attachments,
/images, /forumdata, /include, /ipdata, /templates, /plugins (see in the
options to disable this)
17:59:47 Info: Note: due to forum.jadekwan.net remote robots.txt rules, links
begining with these path will be forbidden: /admin, /attachments, /images,
/forumdata, /include, /ipdata, /templates, /plugins (see in the options to
disable this)
The system seems hanging after 15 minutes of capture. I need to stop and stop
and close the whole HTTrack.
Also, I can see only the first page.
I think the robot.txt rules should be disabled. Am I right?
Also the system hangs while capturing over 15 minutes? Is the robot.txt rules
disabling can solve this? Or do I need to reduce the internal and external
depth of capture to solve such hanging and one-page-only mirror?
In the mean time, i have downloaded the Turn-Flash freeware, do i need to use
it?
Many thanks for any great ideas.
| |