hakluke / hakrawler

Simple, fast web crawler designed for easy, quick discovery of endpoints and assets within a web application
https://hakluke.com
GNU General Public License v3.0
4.41k stars 483 forks source link

No output, process exiting normally #129

Closed zPrototype closed 2 years ago

zPrototype commented 2 years ago

I've installed hakrawler via the provided go install command on my ubuntu:20.04 machine. I've got a list of subdomains separated by line breaks, now when I try to run hakrawler I get not output. The command I ran is: cat probe_again.txt | hakrawler -t 1 -u -d 3 -h "User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:100.0) Gecko/20100101 Firefox/100.0" | tee crawl.txt

I've also tried to add the -subs flag and ran it in docker with: cat probe_again.txt | docker run --rm -i hakrawler_docker -t 1 -u -d 3 -h "User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:100.0) Gecko/20100101 Firefox/100.0" Again the same issue. Running just hakrawler provides the normal help message with all commands as expected.

hakluke commented 2 years ago

what's in probe_again.txt?

zPrototype commented 2 years ago

Subdomains separated by line brake. Like: abc.test.com this.example.com my.cooldomain.net Etc..

garlic0x1 commented 2 years ago

You need to provide full URLs, like https://www.google.com not just the domain name.
Try something like this

https://abc.test.com
http://example.com
hakluke commented 2 years ago

^ what he said

zPrototype commented 2 years ago

Can confirm, that was the issue. Maybe add a check to see if domains are missing a protocol and throw an exception? Not too good with GO otherwise I'd do a PR but I'm sure it would be helpful!

garlic0x1 commented 2 years ago

Can confirm, that was the issue. Maybe add a check to see if domains are missing a protocol and throw an exception? Not too good with GO otherwise I'd do a PR but I'm sure it would be helpful!

wouldn't want to stop the program for one invalid url in a big list, but perhaps there could be a message on stderr