| > > Arfffffffffff .. you are talking about the link's
> are
> > javascript's so then httracker can't understand
> > them !? :0/ arfffff what a job it's going to be
> saving
> > all those file's manually ......
The only solution for 'complex' javascript routines
(that is, routines which dynamically load images using
string expressions) is to add in the source final URLs
which will be detected by the engine. For example:
preload1=new Image();
preload1.src=fooimages/foo.gif';
preload2=new Image();
preload2.src='barimages/foo2.gif';
Of course the webmaster will have to modify its
sources so that it can work.
But remember that javascript is also sometimes used to
avoid offline browsers, and NO offline browser is able
to parse complex javascript code. Even if you merge
the Mozilla code into a spider engine (which may
generate a 10MB program..) there will be other
problems, such as interacting with the user (a mouse
click, wrapped by javascript, can be used to decypher
an URL)
Therefore, if webmasters create websites hostile to
spiders, there is no solution - even for indexing them
(how many search engines are crawling into javascript
and/or java?...)
| |