| | Hi. As mentioned above, I am trying to download a website using httrack website
copier (for example, this site www.asite.com), but I am encountering some
problems. These problems are mainly as follows:
I'm trying to configure the settings to download only sublinks starting with
www.asite.com (e.g., subdomains like www.asite.com/any/sub/link). However, I
keep seeing another site in the list of links being downloaded (e.g.,
www.anothersite.com is also being scanned and attempted to download). Or, the
YT link on the www.asite.com site is also counted as a sublink and is being
attempted to be downloaded.
Media (photos, videos, GIFs, etc.) on the www.asite.com site may be pulled
from another site (for example, photos on the www.asite.com/any/sub/link link
may be pulled from www.image.com). When I set the settings to only pull data
from www.asite.com, these photos are not downloaded.
The sites and media (photos, videos, GIFs, etc.) mentioned above are all made
up. Among the thousands of media files on the site I want to download, there
might be 2 or 3 photos in webp format; no one knows. Assume I'm trying to
download all the content on the site without missing anything.
In summary, I need to configure a setting that will allow me to download
everything from www.asite,com, but not download from other sites, while also
downloading media (photos, videos, GIFs, etc.) pulled from other sites.
If you have a settings file that meets these criteria, I would greatly
appreciate it if you could share it or explain in the comments how I should
customize the settings.
Thank you in advance.
My native language is not English, so I apologize for any spelling mistakes. | |