HTTrack Website Copier
Free software offline browser - FORUM
Subject: Re: Httrack grabbing hudge amount of memory - Linux
Author: Tony
Date: 04/16/2013 21:58
 
> > > Is the -#L1,000,000,000 an issue here? 
> > I have pretty much proven to myself that it was
> the
> > culprit for the memory use. 
> 
> Are you REALLY going to mirror a Billion links? If
> each page averages 50KB, do you REALLY have 50
> terabytes of disk storage to store them on?> 
> It's an issue for you because of you. Get real.

I guess that the issue from a Software Engineering perspective is that it's
really bad news to statically allocate memory unless you *really* need to do
so.  Sure, a billion links is affordable, but that's not the issue.   I've
written many web crawlers and never had to specify the maximum number of links
that would be crawled and doing so can easily incur many penalties such as
poor cache performance.


Tony

 
Reply Create subthread


All articles

Subject Author Date
Httrack grabbing hudge amount of memory - Linux

04/14/2013 15:44
Re: Httrack grabbing hudge amount of memory - Linux

04/14/2013 15:54
Re: Httrack grabbing hudge amount of memory - Linux

04/16/2013 13:47
Re: Httrack grabbing hudge amount of memory - Linux

04/16/2013 15:56
Re: Httrack grabbing hudge amount of memory - Linux

04/16/2013 15:57
Re: Httrack grabbing hudge amount of memory - Linux

04/16/2013 16:13
Re: Httrack grabbing hudge amount of memory - Linux

04/16/2013 21:58
Re: Httrack grabbing hudge amount of memory - Linux

04/21/2013 20:23
Re: Httrack grabbing hudge amount of memory - Linux

04/28/2013 09:50




d

Created with FORUM 2.0.11