This repository houses a scraping engine for the UCPD's Incident Report webpage. The data is stored on Google Cloud Platform's Datastore and ran using Heroku's Dyno functionality.
I'd like to thank @kdumais111 and @FedericoDM for their incredible help in getting the scraping architecture in place. As well as @ehabich for adding a bit of testing validation to the project. Thanks, y'all! <3
^3.11
CENSUS_API_KEY
service_account.json
file stored in the environment
variable: GOOGLE_APPLICATION_CREDENTIALS
GOOGLE_CLOUD_PROJECT
GOOGLE_MAPS_API_KEY
GOOGLE_DRIVE_FOLDER_ID
poetry add [module]
command.
poetry add black
make lint
: Runspre-commit
on the codebase.make seed
: Save incidents starting from January 1st of 2011 and continuing until today.make update
: Save incidents starting from the most recently saved incident until today.make build-model
: Build a predictive XGBoost model based off of locally saved incident data and save it in the data
folder.make categorize
: Categorize stored, 'Information' labeled incidents using the locally saved predictive model.make download
: Download all incidents into a locally stored file titled incident_dump.csv
.