HBase-Writer is a java extension to the Heritrix open source crawler. Heritrix is written by the Internet Archive and HBase Writer enables Heritrix to store crawled content directly into HBase tables running on the Hadoop Distributed FileSystem?. By default, HBase-Writer writes crawled url content into an HBase table as individual records or "rowkeys". Each fetched url is represented by a "rowkey" in an HBaase table. However, HBase-Writer can easily be extended for custom behavior, like writing to multiple tables or anything else. In turn, these HBase tables are directly supported by the MapReduce? framework via Hadoop. HBase-Writer's goal is to facilitate in fast large distributed crawls using Heritrix and to save and manage Web-scale content using HBase.
=What steps will reproduce the problem?=
1. Crawl a small site.
2. Copy the crawl to a new job and Set the option of only_new_records to "true"
3. Crawl the site again
If you monitor the wire, the content is fetched. the content should not be
fetched if set to true.
The fix is to move the only_new_records if block to shouldWrite() instead
of shouldProcess().
Original issue reported on code.google.com by ryan.justin.smith@gmail.com on 13 Feb 2009 at 12:03
Logic has been moved to shouldWrite() and tested further. Tested on a few small
crawls and heritrix is not downloading any content that has been fetched into
hbase
already.
Original comment by ryan.justin.smith@gmail.com on 16 Feb 2009 at 6:56
Original issue reported on code.google.com by
ryan.justin.smith@gmail.com
on 13 Feb 2009 at 12:03