datopian / aircan

💨🥫 A Data Factory system for running data processing pipelines built on AirFlow and tailored to CKAN. Includes evolution of DataPusher and Xloader for loading data to DataStore.
https://tech.datopian.com/flows/
Other
29 stars 6 forks source link

Analysis for Logging system #2

Open rufuspollock opened 4 years ago

rufuspollock commented 4 years ago

Logging and reporting is a crucial aspect of a data factory system like this.

Job stories

When a Run is initiated by an Operator they want to see it is running and be notified of application and (meta)data errors as soon as possible, especially “halts” so that they can debug and re-run

If there are a lot of (data) errors I want to examine these in a system that allows me to analyse / view easily (i.e. don’t have my page crash as it tries to load 100k error messages) I don’t want to receive 100k error emails …

When a scheduled Run happens as an Operator (Sysadmin), I want to be notified afterwards (with a report?) if something went wrong, so that I can do something about it …

When I need to report to my colleagues about the Harvesting system I want to get an overall report of how it is going, how many datasets are harvested etc so that I can tell them

Domain Model

Status info: this is Run is running, it is finished, it took this long …

(Raw) Log information …

Reports / Summaries e.g. 200 records processed, 5 errors, 2 warnings, 8 new dataset, 192 existing records updated

4 cases

hannelita commented 4 years ago

Google cloud composer already provides a lot of logs. We may be able to create a sink on GCP Operations Logging and redirect the created logs to another service