HTTrack Website Copier
Free software offline browser - FORUM
Subject: HTTrack will not copy local webpages or this URL:
Author: Dapianoman
Date: 04/17/2003 05:48
 
http://www.angelfire.com/pq/paradisegardens/cannabisholyanoi
ntingoil.html

is a webpage I just created from an earlier version of the 
document, which had the <u> underline html instead of <a 
href> linking code. 

I can't seem to get HTTrack to copy the pages linked from 
the URLs from the page...!

The version of the document I have was prepared by taking 
the plain text off of the browser (not source, but 
displayed text), then inserting it into the Text2HTML 
conversion webpage. It did a very good job of creating a 
webpage with active links, but for some reason HTTRACK will 
not copy all the URLs... and I KNOW how to use this program 
well... I set it to follow robots.text, then tried not 
following robots.txt, set the client to MSIE 6.0 so the 
Angelfire server won't balk, set it to go both up and down, 
get html files first, get all links to a file, get all 
filetypes I'm looking for, defaul site-structure, default 
on tolerant requests for servers, parse javascript for 
urls, etc.... 

I also discovered that HTTrack doesn't know what to do with 
web pages it doesn't have to download; i.e. HTML web pages 
saved onto one's one hard drive! 

It would be great to be able to create a web page with 
links, and use HTTrack to check all the links to make sure 
they are all current and active ! *without having to first 
upload the webpage to a server on the Web*

can you help??
TIA, 

Dapianoman

 
Reply


All articles

Subject Author Date
HTTrack will not copy local webpages or this URL:

04/17/2003 05:48
Re: HTTrack will not copy local webpages or this URL:

04/17/2003 06:55
Re: HTTrack will not copy local webpages or this URL:

04/17/2003 23:57




7

Created with FORUM 2.0.11