| Hi, I already search over the forum, this is specific, sorry if this is a noob
question it is not clear for me ;)
I understand that Filters (priority, order) is from first to bottom, but I
don't know if this admits 'logic' or 'reasoning', as explicit 'exclude' /
'include' and exceptions (overriding)
for example
I want to archive a site (example is fictional site)
domain is cg3dtutorials.tk
I want some specific pages:
(mime:text/html) (not all html) only on defined patterns & mime:video/*
+mime:video/*
+mime:text/html
-cg3dtutorials.tk/*
+cg3dtutorials.tk/tag/*
+cg3dtutorials.tk/video/*
+cg3dtutorials.tk/watch/*
+cg3dtutorials.tk/category/*
+cg3dtutorials.tk/thread/*
+cg3dtutorials.tk/new/page/*
... but only mime:text/html & mime/video/* defined by inclusions (+)
I put -cg3dtutorials.tk/* as a general explicit exclusion, then explicit
inclusions.
I want to know if the desired result is (denied - except +) or if this will
denied all (as the first is a general exclusion)
The desired-result is
+html for every link with desired pattern 'only', excluding all the other
pages.
I assume the 'crawler' will follow only desired links and not the rest, is
this right?
+ all videos with the desired pattern
The filters are only separated by line or I can combine filters in one line as
operators?
For example
+cg3dtutorials.tk/tag/* +mime:text/html
is not the same than
+cg3dtutorials.tk/tag/*
+mime:text/html
or it is the same?
For every link the crawler checks all filters or only the first (denied or
accepted) ?
| |