istresearch / scrapy-cluster

This Scrapy project uses Redis and Kafka to create a distributed on demand scraping cluster.
http://scrapy-cluster.readthedocs.io/
MIT License
1.17k stars 323 forks source link

Crawler unhandled exceptions not logged #259

Closed getorca closed 3 years ago

getorca commented 3 years ago

With only LOG_ENABLED = false in settings.py , default behaviour, when running with docker compose you loose logging of unhanded exceptions such as KeyErrors. This is in the crawler service.

expected behaviour: unhanded exceptions get logged

It might make sense to have:

LOG_ENABLED = str2bool(os.getenv('LOG_ENABLED', True))
LOG_LEVEL = os.getenv('LOG_LEVEL', 'ERROR')

as the default so that unhandled exceptions are still logged and visible.

madisonb commented 3 years ago

This is a Scrapy setting, not a Scrapy Cluster setting. This project has no control over what that setting does or how it impacts scrapy logging.

Closing.