Closed GoogleCodeExporter closed 9 years ago
Hi Yasser..
Any updates on this ?
Original comment by w3engine...@gmail.com
on 15 Feb 2012 at 10:52
This is by design. Crawler4j is designed for crawling domains and extracting
content, not to detect broken links. This is a specific need for your
application. Of course, you can customize it to support your scenario. For
example, whenever, you see a broken link, keep it somewhere (memory, db, ...)
and whenever a new page is visited go through its links and see if any of the
known broken links is among them.
-Yasser
Original comment by ganjisaffar@gmail.com
on 17 Feb 2012 at 5:12
Original issue reported on code.google.com by
w3engine...@gmail.com
on 5 Feb 2012 at 4:49