18F / identity-analytics-etl

ETL and SQL scripts for Login.gov data warehouse and business intelligence
10 stars 6 forks source link

Automate The ETL Pipeline Status Check Up #181

Closed MacHu-GWU closed 5 years ago

MacHu-GWU commented 5 years ago

User story

As the analytics ETL pipeline developer, I would like to automate the status / data healthy checkup, so that I don't have to log in to AWS Console / jumphost to manually run some script over and over again.

The manual check up:

  1. SSH to Jumphost EC2 instance, manually run some query to identify number of new ingested data.
  2. Login to AWS Console, visually check some invocations with errors, and go to Cloudwatch to identify whether it is normal (Time Consuming).

Notes

What is the value to the user in this story?

The business persons are periodically checking the QuickSight dashboard. They found the Pipeline Outage on Sep 25, 2018 and notified analytics team.

Since there is no user periodically checking QuickSight right now, and there's no easy way to allow non developer checking the ETL pipeline status. It is really hard to know that if there's an outage when @MacHu-GWU is out.

What are things we should consider when making this story

Acceptance Criteria

Tasks to complete the story

Definition of Done

lauraGgit commented 5 years ago

@MacHu-GWU it seems like we should wait until the new pipeline is implemented. What do you think?

MacHu-GWU commented 5 years ago

@lauraGgit

There are two parts to check for Pipeline Status:

  1. Check Redshift Data
  2. Check Pipeline Invoke Stats, duration, error rate.

I think <1> will not change in the new solution. We can still do <1>.

lauraGgit commented 5 years ago

migrated to jira