HTTrack Website Copier
Free software offline browser - FORUM
Subject: Get only urls
Author: Sandeep
Date: 09/08/2017 14:49
 
Requirement - For a given url (e.g: <https://mydomain.net>), crawl through the
url and show only urls (both external and internal) and later get the
interested response header for each url. 

For above scenario, I tried 'httrack --spider <https://mydomain.net> but I
don't see the list of urls. Also tried '--external' but it only shows the
links on the fly but don't list or store it. 

Am I missing something? I am quite new to this field. 
 
Reply


All articles

Subject Author Date
Get only urls 09/08/2017 14:49




1

Created with FORUM 2.0.11