-
Right now the Scrapy Spider Workflow #7 gets trigger on every push (pr and merge) it should only work with pr's.
-
### Brand name
Huffer
### Wikidata ID
Q107862765
### Store finder url(s)
https://www.wikidata.org/wiki/Q107862765
https://www.wikidata.org/wiki/Special:EntityData/Q107862765.json
### Store fi…
-
### Steps To Reproduce
Steps to reproduce the behavior:
1. build *python312Packages.scrapy*
### Build log
Fails during the test phase:
```
=================================== FAILURES ======…
-
I have a spider that after a certain time begins receiving 503 Service Unavailable errors. If given enough download delay, I can avoid these errors. The autothrottle documentation has led me to believ…
-
The memory backends are all implemented using heapq. This allows for some succinct code when using different crawl ordering, but it's less efficient than choosing more appropriate datastructures for e…
-
Fetched 2 brands/shop/pest_control from NSI
Missing by wikidata: 1
### Brand name
Truly Nolen
pest control, termite control and exterminator
### Wikidata ID
Q7847671
https://www.wikidat…
-
There seems to be very little documentation on catching exceptions with scrapy. But before I open issue(s?) around that, I wanted to check.
My code has a `raise MemoryError()` which correctly trigg…
-
I have a Splash running in the docker, beside the issue described in the https://github.com/scrapinghub/splash/issues/586, I found Splash will crash after running for a little while.
> May 1 14:21…
-
Teresina-PI is not collecting gazettes for weeks.
"It works in my machine" :rofl: so one guess is that they are blocking access of our spider that is running in Scrapy Cloud datacenters (located in…
-
The handler is not allowing enough time for the new browser to launch after a crash.
Sample spider adapted from #167.
```python
# crash.py
import os
from signal import SIGKILL
import psuti…