| First off, I have been using WinHTTrack for almost two
years now. It was amazing software when I first started
using it and has had several out of this world
improvements since. Thanks Xavier Roche et al.
Just a little background... I am working for a non-profit
group called the WiderNet Project (www.widernet.org) and
one of our projects is building digital libraries and
distributing them in developing regions of Africa.
We have been obtaining permissions from various web
authors and spidering their websites to place on 120 GB
hard drives and delivering them to universities in Africa
with no or little Internet connectivity. Many times we
have made libraries at universities in Africa hundreds of
times bigger with one 3.5" drive. Eventually we will be
updating our hard drives with Satellite forward only
data.
We currently have about a hundred different sites and we
are growing.
Q: Can we update all these sites (different project
files) at once with one script?
Q: Can we tell WinHTTrack to store cache and settings for
an individual site in a different place then the actual
copied site (for example: two seperate drives on the same
machine) and still have it update the site? (would have
one master hard drive with all sites on it and make direct
copies of it for sending to Africa, no need to send
overhead files from HTTrack to Africa.)
Q: If a site we are downloading new links to a site we
have already downloaded can we tell HTTrack to
automatically link to our local copy (seperate project
files)?
Q: We have downloaded www.bartelby.com and each page has
a Java Script that points to an outside link (advertising
server). This can do some weird things when there is not
Internet connectino present and even weirder things when
there is a slow connection. We have already had limited
success with programs that can strip out this code after
the download. Is there a pernament solution HTTrack can
do during the download so we don't have to fix it after
every update?
We'd also love to hear from any hardcore proxy server
programmers. It would be great to have these sites work
seamlessly on campuses in Africa so that students can type
in the actual URL of a site and either get the version
cached on a 120GB hard drive with other entire sites or
the new page if it has been updated (possibly depending on
rights/status (student, staff, or faculty etc.)).
Sorry so long. Wanted to let the developers know that
their work makes a difference globally. Not only a
difference but it has changed lives and sparked many tears
of joy!
Thanks,
mike
(feel free to reply with any comments to my email address
if you wish)
| |