-
I see that each site has a .db file (sqllite) that contains all the URLs and some basic metadata. My question is, how is that collected and put into these .db files. I would love to see the script tha…
-
I've been using Cyberdrop-dl for a few months but recently I tried to do a batch job of 50 links from bunkr. The batch job runs the problem is that there are a lot of files that cannot be downloaded f…
-
## Feature Request
We do not know, if the scrapes data of `[[input.prometheus]]` are fetched (or missed).
### Proposal:
### Current behavior:
Missing data, no data if scraped.
### Desir…
-
Allow users to search albums/artist (and maybe songs) that share a domain as an external link.
This would allow users to find which albums are available on booth.pm/bandcamp and find which artists …
-
- **Agency Slug:** ca_los_angeles_pd
- **Module:** clean/ca/los_angeles_pd.py
- **URL:** https://www.lapdonline.org/senate-bill-1421-senate-bill-16-sb-16/
-
Where is dist/index.js?
fehmi updated
5 months ago
-
I want to automatically archive stories that I've added to my personal bookshelves, however I can't seem to find a proper way to extract story URLs from my own bookshelves.
Using the following comm…
-
Hello, I have some ideas for new projects to start.
1. TikTok scraper tool, scrapes a list of “tubeuppable” URLs from a User, Tag, Effect, Search, Location, or Sound
2. A tool to archive file shari…
-
Hello,
I would like to add an example showcasing how a distributed web-crawler can be made using `Golang` and a `Kafka` cluster on `Kubernetes`. Some ideas:
- A web-extension which collects the …
-
Hi, first of all thank you for the code!
I am however having the problem that when scraping multiple pages of reviews for the same product, only the first page gets scraped. The other pages get "sc…