-
Sur demande je peux fournir le code PERL du moteur d'exportation d'etaamb. L'idée est que ce dernier soit remplacé par un composant plus moderne, en python, et Open Source.
-
Great tool, but I'm struggling with few things.
How actually some of the inner workings operate? I want to calculate hash without running scraper, but can't the md5,sha and others do not match your…
-
Thanks again for the Waltham Forest Hackathon last night! Here's some notes / code form our team for going forward:
* requirements include scraping events from a whitelist of sites, however a) t…
-
`lxml` has an html5parser that can handle some of the inanities that bad HTML pages present.
For example, this page:
http://media.ca11.uscourts.gov/opinions/unpub/logname.php?begin=9720&num=485&num…
-
Using former version (7) I could convert the HTML file generated from MS WORD 2010 without any issue. However after I upgraded to latest version (8), it doesn't work any longer. It's identified as cor…
-
## Bug report
Failure to scrape Series on Systems with MySQL Database
### Describe the bug
This behaviour was observed on Systems working with MySQL Database: when trying to scrape some TV seri…
-
Usecase: Trying to compare these three mediums.
- https://tvtropes.org/pmwiki/pmwiki.php/WebVideo/Jreg and https://tvtropes.org/pmwiki/pmwiki.php/Characters/Jreg
- https://tvtropes.org/pmwiki/pmwiki…
-
Hi,
I want to download comments for some posts (either by post_id or post_url). I was doing:
```
next(get_posts(post_urls=[post["post_id"]],
cookies=COOCKIE_FII…
ekote updated
3 years ago
-
First of all thank you for your great script.
I tried the demo and learn the syntax from wiki to extract the content from html to be RSS.
However I think it syntax is really difficult to understand.…
-
*Please use the following format for a ZIM creation request (and delete unnecessary information)*
- Website URL:
- License: **CC BY-NC-SA 3.0**
- Desired ZIM Title: **Minecraft Wiki (zh)**
- De…