stacktonic-com / gcp-export-bigquery-to-firestore

Export BigQuery data to a Google Firestore database collection, using a Google Workflow and Cloud Function.
https://stacktonic.com
5 stars 1 forks source link

Possible optimisations for large amount of records #1

Open MarkEdmondson1234 opened 2 years ago

MarkEdmondson1234 commented 2 years ago

Many thanks for this and I look forward to the blog post :)

I have just seen your slides, so perhaps you covered this, but I have a case with 2.5million records in the BigQuery table that I think will cause out of memory issues. That solution is using Dataiku as its within an existing setup, but I'd like it to be able to work with Worflows too. For that I think possible modifications would be:

What do you think?

MarkEdmondson1234 commented 2 years ago

Ah found this - paging through BigQuery https://cloud.google.com/workflows/docs/paginate-bigquery-results

DavidHemmerle commented 2 years ago

Hey @MarkEdmondson1234, it would be great to have a look at your solution of this problem! Have you already come up with an approach to apply the paging through BQ to this export to Firestore for bigger amounts of data?