Closed apsaltis closed 8 months ago
Hi, the crawler only crawls urls that exist, there is no decoding done on the urls. If you have a lot of operations going I would not use println!
and change it to stdout lock. I would also avoid using scrape on a website that large for cprime.com. Scrape stores the html content throughout the crawl. I think the issue could have been due to memory constraints if the content does not exist on the website.
There was actually an issue with the semaphore on the scrape calls with chrome leading to memory issues. A fix is coming out.
Should be fixed in 1.85.4
, thanks for the issue!
Thanks for the info and also the pointer on println! -- greatly appreciated.
Hi, I have the following code: