Liking cljdoc? Tell your friends :D

clj-robots-parser.core


is-crawlable?clj/s

(is-crawlable? robots-txt url user-agent)

Does the given parsed robots.txt permit the given URL to be crawled by the given user-agent?

Does the given parsed robots.txt permit the given URL to be crawled
by the given user-agent?
sourceraw docstring

parseclj/s

(parse content)

Parses the given string (content of a robots.txt file) into data that can be queried.

Parses the given string (content of a robots.txt file) into data that
can be queried.
sourceraw docstring

query-crawlableclj/s

(query-crawlable {:keys [agent-groups]} url user-agent)

Determines whether and explains why the given parsed robots.txt does or does not permit the given URL to be crawled by the given user-agent.

Determines whether and explains why the given parsed robots.txt does
or does not permit the given URL to be crawled by the given
user-agent.
sourceraw docstring

robots-txt-lineclj/s

source

stringify-query-resultclj/s

(stringify-query-result {:keys [raw-content]}
                        {:keys [because]}
                        &
                        {:keys [context] :or {context 1}})

Creates a user-readable string explanation of a query-crawlable result by providing contextual highlighting of the source robots.txt that produced it.

Creates a user-readable string explanation of a query-crawlable
result by providing contextual highlighting of the source robots.txt
that produced it.
sourceraw docstring

cljdoc is a website building & hosting documentation for Clojure/Script libraries

× close