Crawly.Middlewares.RobotsTxt (Crawly v0.17.2) View Source
Obey robots.txt
A robots.txt file tells search engine crawlers which pages or files the crawler can or can't request from your site. This is used mainly to avoid overloading a site with requests!
No options are required for this middleware. Any tuple-based configurations options passed will be ignored.
Example Declaration
middlewares: [
Crawly.Middlewares.RobotsTxt
]
Link to this section Summary
Functions
Callback implementation for Crawly.Pipeline.run/3
.
Link to this section Functions
Callback implementation for Crawly.Pipeline.run/3
.