-
I've noticed that in the loop to create `ScrapeWorkerFactory`s in `core.py`, there's a line that loops though every proxy in the given file (if one chooses to use one), which ends up creating more bro…
-
Maybe a cron job that runs nightly, to replace the data...
However, the dataset only includes agencies that have current listings, and there is a hard-coded lookup to map the agencyname from the da…
-
#### What we want:
- [x] a dynamic scraper function that takes a list of `blogs` as input, cycles through them, and scrapes them.
- [x] fetch all media from the post with img/video credits and add…
-
## Issue
**Asynchronous processing using async function**
The submission has the right idea to handle the scraping process asynchronously
https://github.com/tmphu/nimble-be/blob/395b608063698ee51…
-
I have a recurring job that runs a "check" on a certain object. I enqueue a job with a specific job id, built from a job key and the object's UUID, such as `check_object_{object_uuid}`.
Then, when …
-
Tech
-
Finance
- Moneycontrol [https://github.com/Web-Scrapers/Money-Control](https://github.com/Web-Scrapers/Money-Control) also they might have a api aswell cant find clear doc though
- CNB…
-
Hi,
when starting, it gets this endless loop, and I can't get any subscriptions:
![image](https://github.com/DIGITALCRIMINALS/UltimaScraper/assets/81478245/76ed5c8e-2e99-46a7-810e-2760a7f06ed8)
…
-
I tried the company filter. It works well for the most part, but there seems to be an issue with the program - It only returns the first 25 jobs while there are hundreds on the website. I added someth…
-
when cache file larger than 10,000, parser serp from caching is very very slow.
for 12000 cache, it probably need to spend 2 hours.
``` python
for path in files:
print('processing {num_cach…
-
when running the example am getting the above error how do I overcome the error ?