Closed forgeries closed 2 years ago
You can delete the data directly using the redis CLI.
Close this as solved
For a more automatic solution, you could use a scrapy extension. Unfortunately, I can't share our exact solution, but its likely a generic extension could be implemented and provided by the base library.
At a high level:
SCHEDULER
is set to scrapy_redis.scheduler.Scheduler
How to delete duplicate files when the crawler task is all over.