In #731 discussion, some smart person noticed that if you change bunkrr.su links so that they start with www. apparently it helps circumvent DDoS-Guard.
e.g.
original (failing) link: https://bunkrr.su/a/<bla>
I found four occasions in my recent logs where scraping bunkr was repeatedly failing with DDoS-Guard to no avail. (Tried it multiple times, it's always failing. Some other bunkr links were ok. I have no idea what's wrong with those particular links.)
Verified that I can access those links with a browser with no problem. (Importing DDoS-Guard cookies from browser does not help in this particular case.)
Created a temporary url.txt with those four links.
Used this url.txt with CDL and observed the same DDoS-Guard failures during scrape.
Edited the url.txt and prepended each link with www..
No more DDoS-guard, and download succeeded.
Sounds like an easy fix. Except that those links come from a forum thread. And of course, I cannot go to the forum and ask people to edit their links.
Moreover, I have no idea for how long it will keep working.
But maybe add a feature to CDL that if a link fails without www. then add www. and retry? Maybe it will sometimes help not only with bunkr but other services, too. And maybe even removing www. will help in some similar cases.
In #731 discussion, some smart person noticed that if you change
bunkrr.su
links so that they start withwww.
apparently it helps circumvent DDoS-Guard.e.g.
https://bunkrr.su/a/<bla>
https://www.bunkrr.su/a/<bla>
I tried it and—surprisingly—it works.
url.txt
with those four links.url.txt
with CDL and observed the same DDoS-Guard failures during scrape.url.txt
and prepended each link withwww.
.Sounds like an easy fix. Except that those links come from a forum thread. And of course, I cannot go to the forum and ask people to edit their links.
Moreover, I have no idea for how long it will keep working.
But maybe add a feature to CDL that if a link fails without
www.
then addwww.
and retry? Maybe it will sometimes help not only with bunkr but other services, too. And maybe even removingwww.
will help in some similar cases.