Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add "revert to cached" load action #28

Open
hamishmorgan opened this issue Apr 26, 2015 · 0 comments
Open

Add "revert to cached" load action #28

hamishmorgan opened this issue Apr 26, 2015 · 0 comments

Comments

@hamishmorgan
Copy link
Contributor

Currently, when a robots.txt is retrieved there are three possible outcomes:

  1. Fully disallow the site (e.g after 403 "Forbidden" HTTP response.)
  2. Fully allow the site (e.g after 404 "Not Found")
  3. Conditionally allow, based on the result of parsing the response.

Sometime, however, it would seem more sensible to back-off to a previously cached response, when that option is available. Examples where this behaviour would be desirable include:

  • Rate limiting responses; such as 420 "Enhance your calm", 429 "Too Many Requests", 509 "Bandwidth Limit Exceeded", and 598/8 "Netowork timeout error"
  • Cache-control responses: 304 "Not Modified"
  • Temporary errors: 408 "Request Timeout"

This would nice to have, but it's probably very important. I imagine these conditions don't occur very often.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant