scrapy / scrapy

Scrapy, a fast high-level web crawling & scraping framework for Python.
https://scrapy.org
BSD 3-Clause "New" or "Revised" License
52.17k stars 10.44k forks source link

Refactor signals #8

Open dangra opened 12 years ago

dangra commented 12 years ago

Refactor signal handling similar how Django did for 1.0.

According to them, the new approach brings up to 90% speed improvements

pablohoffman commented 11 years ago

Signals got their own API on 0.15 (no longer need to import pydispatcher) but we can still rewrite the backend (which is what this issue is for) - it will actually be easier to do it now.

artemdevel commented 11 years ago

Perhaps we still need some refactoring here (or maybe I found a bug). I added some test code here https://gist.github.com/artem-dev/4996685 So at first I tried to use new API as shown in test5.py but this approach didn't work for me. I managed to make it work as shown in test5_2.py but as for me this is even less convenient than old approach which is shown in test5_1.py Did I use this new API wrong way or is it some sort of bug?

artemdevel commented 11 years ago

Just after my previous post I decided to do one more small test, looks like this change https://gist.github.com/artem-dev/4996755 fixes the issue I described earlier, so sm = SignalManager() is working, but I'm not sure if using dispatcher.Any instead of dispatcher.Anonymous is ok for the rest of the engine (but my simple example worked fine)

artemdevel commented 11 years ago

hi @pablohoffman this code https://gist.github.com/artem-dev/5046355 is working, is it a correct way to use the new SignalManager?

pablohoffman commented 11 years ago

I'm pasting a comment I wrongly made in another issue:

@artem-dev you are instantiating a SignalManager in your own spider, that's not how it's supposed to work. You should be connecting to the SignalManager of the Crawler controlling your spider (accessible through crawler.signals).

User code should never instantiate a SignalManager, but use the one instantiated by framework code (and accessible through the Crawler object).

I hope this is more clear now.

pablohoffman commented 11 years ago

That's right @artem-dev. However, doing it in parse would cause it to be called many times, so start_requests would be more appropriate.

artemdevel commented 11 years ago

yeah, it was just for a test

kmike commented 10 years ago

For the reference, django's implementation: https://github.com/django/django/tree/master/django/dispatch One advantage of switching to django implementation is that is supports Python 3.x. But I haven't checked if it is possible to switch.

jdemaeyer commented 9 years ago

Here is a partial proof of concept for switching the backend to django.dispatch while conserving support for the old signaling API: http://jakobdemaeyer.com/x/scrapy_signaling_proof.html

In this implementation, the SignalManager is deprecated and replaced by a NewSignalManager whose sole purpose is backwards-compatibility. The major open question is how to assign the Crawler instance as default sender for a signal, as the signals are now essentially completely decoupled from anything else.

A different implementation could preserve using the SignalManager by solely rewriting the backend, using the NewSignalManager implementation without warnings. Signals would then still be emitted (and handlers registered) via the Crawler.signals attribute instead of on the Signals themselves.

jdemaeyer commented 9 years ago

Here's a list of signaling frameworks that might be of interest: http://www.fortpedro.com/blog/2015/1/observer-pattern-event-oriented-programming-in-python

kmike commented 9 years ago

another option: https://pypi.python.org/pypi/blinker

eLRuLL commented 9 years ago

I would like to work on this issue :smile:

curita commented 9 years ago

If anybody is currently working on this issue let us know, otherwise @eLRuLL: all yours!

kmike commented 9 years ago

I've profiled a simple spider which downloads a page and follows all links from it using LinkExractor. There is a non-standard CrawlerProcess which listens to all signals from its Crawlers and re-sends them through its own SignalManager to provide a single place to listen for all signals - nothing complicated. I haven't checked it with a standard CrawlerProcess.

For this spider signal handling (SignalManager.send_catch_log function) is one of the bottlenecks - it takes ~5 times more time than HTML parsing.

redapple commented 7 years ago

See https://github.com/scrapy/scrapy/pull/2030