HTTrack Website Copier
Free software offline browser - FORUM
Subject: Re: Bug in the way httack handles multiple URLS
Author: Xavier Roche
Date: 12/13/2002 22:43
 
> URL's specified in the project:
> www.geocities.com/alpha
> www.geocities.com/beta
> www.geocities.com/charlie
> httrack now tries to get *everything* from geocities, not 
> just the urls specified and below.

First are you sure these are the correct URLS? Try these 
ones:
www.geocities.com/alpha/
www.geocities.com/beta/
www.geocities.com/charlie/

I suspect www.geocities.com/alpha to be redirected to 
www.geocities.com/alpha/, which slows down a bit the 
capture, and is causing the problems you are having


> The only workaround I have is to create a project for 
each 
> URL, which then means things operate as you'd expect and 
> external stuff outside of the url or below is not grabbed.

Another solution is to use filters: (Options/Scan rules)

-www.geocities.com +www.geocities.com/alpha/* 
+www.geocities.com/beta/* +www.geocities.com/charlie/*
 
Reply Create subthread


All articles

Subject Author Date
Bug in the way httack handles multiple URLS

12/13/2002 22:13
Re: Bug in the way httack handles multiple URLS

12/13/2002 22:43
Re: Bug in the way httack handles multiple URLS

12/13/2002 22:44
Re: Bug in the way httack handles multiple URLS

12/13/2002 23:32
Re: Bug in the way httack handles multiple URLS

12/14/2002 08:33
Re: Bug in the way httack handles multiple URLS

12/14/2002 22:41
Re: Bug in the way httack handles multiple URLS

05/07/2005 19:45




f

Created with FORUM 2.0.11