Closed IanRFerguson closed 10 months ago
Ok one issue to address @Jason94 ... for "really big files" (e.g., the New York Catalist voterfile) I'm getting a timeout trying to stream the data into GCS (requests.exceptions.ConnectionError: ('Connection aborted.', timeout('The write operation timed out'))
)
Do you think it's worth trying to handle this (either by checking the size of the file or with a try / catch), or should we just stick with the local file -> blob method that exists in the current function?
Ok one issue to address @Jason94 ... for "really big files" (e.g., the New York Catalist voterfile) I'm getting a timeout trying to stream the data into GCS (
requests.exceptions.ConnectionError: ('Connection aborted.', timeout('The write operation timed out'))
)Do you think it's worth trying to handle this (either by checking the size of the file or with a try / catch), or should we just stick with the local file -> blob method that exists in the current function?
There IS a timeout
param in the Google API, maybe I'll just mess with that
The current
gcs.unzip_blob()
implementation only handlesgzip
files, we're updating to handle other forms of compression