Closed b5 closed 5 years ago
hi @b5, do you mean "are we doing a dupe check" (no), or "are we purposely adding large numbers of duplicate urls via the extension (also no). Am nhoping the 2nd is not something that's happening
As long as we use a spreadsheet backend, querying our storage medium for the URL will be difficult/slow. We already check if it's been archived by archive.org. At some point we should probalby rewrite the extension so that it integrates directly into archivers 2.0, yes? But probably we want to figure out the schope of future archiving events/activities a bit better first?
A while back @sonalranjit was trying to move away from spreadsheets over to adding URL's directly archivers, but at the time it was a bit of a moving target, is it time to move over to arhcivers 2.0? If so, give some skeletal guidance, maybe we can do it tonight.
@b5 would be too much to maintain if we just create a standlone db service for the extension submissions or submitting it to patchbay?
api.archivers.space/v0/uncrawlables @mi-lee
changed title to reflect new direction!
This issue has been automatically marked as stale because it has not had recent activity. It will be closed in seven days if no further activity occurs. If it should not be closed, please comment! Thank you for your contributions.
This is safe to close, I think, as the new new direction is to integrate with walk once it's ready.
Hi team! I've been working on integrating the EOT tool output into archivers 2.0, just wanted to check to see if the behavior of adding identical urls is intended.