In the case of rerunning a inprogress job, it'll "forget" about already visited sites and start overwriting existing rows and crawling everything again. We should update the spider so that it checks the database to see if a site has already been visited before vistiting it again.
In the case of rerunning a inprogress job, it'll "forget" about already visited sites and start overwriting existing rows and crawling everything again. We should update the spider so that it checks the database to see if a site has already been visited before vistiting it again.