Whole crawler dies because "failed to lookup address information: Name or service not known"
I am not able to reproduce it in a simple example (it may be a transient error), but I have gotten this error regularly and it kills the crawler completely.
This is my crawler:
I am on version 1.0.4 and I was crawling crawlee.dev (though it doesn't fail in a specific page)
5 Replies
I think it is related to the new release because I had not seen this error before upgrading to 1.0.4 (from 1.0.3)
This error specifically shows up while Crawlee tries to download the robots.txt file.
you can try to
thanks, I tried and you are right, the error doesn't appear. I would like to respect the robots.txt though...
Thank you for bringing this to our attention. This is an error, and we will endeavor to correct it in the next release.
fix is on the way https://github.com/apify/crawlee-python/pull/1524
GitHub
fix: Improve error handling for
RobotsTxtFile.load by Mantisus ·...Description
This PR adds error handling for RobotsTxtFile.load. This prevents crawler failures related to network errors, DNS errors for non-existent domains (e.g., https://placeholder.com/), or u...