HTTrack Website Copier
Free software offline browser - FORUM
Subject: Re: Downloading an endless scrolling tumblr page
Author: fictor
Date: 08/24/2012 18:32
 
All before, WARNING!!
th3-rest-is-still-unwritten.tumblr.com/  may contain some nsfw content. Sorry
for that.


OK i will explain better now.
Here is what i have done so far.

Mirroring site
<http://th3-rest-is-still-unwritten.tumblr.com/>

MY Settings
Action: Download website(s)
Get non-html files related to a link(checked)

MY Scan Rules
+*.png +*.gif +*.jpg +*.css +*.js -ad.doubleclick.net/*
-*.tumblr.com/* +*.media.tumblr.com/* -*.media.tumblr.com/avatar_*.png
+*.static.tumblr.com/* +th3-rest-is-still-unwritten.tumblr.com/*
-*photobucket.com/*

The problem is, httrack will only index the main page (page 1). It will not
index page 2, page 3, page... because there no links for that pages on
anywhere. And i cannot do it all manually, in several paged blogs.
How can i solve this problem?

>external depth has been reported broken in the newest versions (fails to
stop)
I ran into this problem before but i thought it is my failure. 
So, what's the latest version which is correctly works with that function?If i
set external depth to 1 (on a previous version), is it gonna save all outgoing
links. And it will not take links from an outgoing link (unless, i set depth
to 2)
Is it right?

PS: I chose that blog because, it has little content and it contains at least
two pages, with auto scrolling script. So, it fits perfectly my purposes and
tests.
 
Reply Create subthread


All articles

Subject Author Date
Downloading an endless scrolling tumblr page

08/22/2012 23:19
Re: Downloading an endless scrolling tumblr page

08/23/2012 17:27
Re: Downloading an endless scrolling tumblr page

08/24/2012 18:32
Re: Downloading an endless scrolling tumblr page

08/25/2012 12:51




5

Created with FORUM 2.0.11