When a remote server is not responding, we have 2 issues.
1) The request for robots.txt takes a long time to fail.
2) When it does fail, requests.exceptions.Timeout is not thrown. Instead we get a requests.exceptions.RequestException, which does not abort the link check. So we still have to wait for the link check to timeout before moving on.
This patch uses the configured timeout value for robots.txt requests, which will throw requests.exceptions.Timeout and abort the current link check if the remote host does not respond after the timeout .
When a remote server is not responding, we have 2 issues.
1) The request for robots.txt takes a long time to fail. 2) When it does fail,
requests.exceptions.Timeout
is not thrown. Instead we get arequests.exceptions.RequestException
, which does not abort the link check. So we still have to wait for the link check to timeout before moving on.This patch uses the configured timeout value for robots.txt requests, which will throw
requests.exceptions.Timeout
and abort the current link check if the remote host does not respond after the timeout .