This repository contains automations for Treetracker processes using an open source worflow orchestration tool called Airflow (https://airflow.apache.org/). Airflow allows us to flexibly and robustly create and schedule jobs to run on a schedule or be manually triggered. Airflow uses the python language and requires python 3 to be installed to run.
To run Python code without installing airflow:
lib
folder.Fork this repo and clone it into your local directory:
Add this to your ~/.bash_profile
then
Modify your airflow.cfg:
This approach seems to work well on MacOS X, but on Windows requires many extra dependencies. More information is available at https://airflow.apache.org/docs/apache-airflow/stable/installation/installing-from-pypi.html#
You may run a ubuntu instace in docker, and install airflow there using pip3. There may be a docker image available on dockerhub for this purpose as well, but we have not tested any publicly available images at this time.
If run airflow in docker, using ubuntu, all pre-requisites for airflow is just: python (2/3), pip and bind the 8080 port to allow locally visit to the admin dashboard of airflow, here is an example of docker command to run the container: docker run -it -d -v ~/temp/airflow/mydata:/mydata -p 8080:8080 --name myairflow ubuntu
Airflow DAGs are authored using any editor that you choose. When you author DAGs in the configured airflow DAGs folder (defaults to /Users/{user}/airflow/dags on MacOS X), airflow detects the automatically and runs them. You can view the outputs of each run in the airflow web panel.
airflow standalone
on the command line, note that this command gives you an address for the airflow web panel as well as the password for the admin userwith DAG(
'reporting-schema-copy', <<<--- this is the DAG id, it cannot be duplicated
default_args=default_args,
description='Calculate earnings for FCC planters',
schedule_interval= '* * * * *',
#schedule_interval= '@hourly',
start_date=datetime(2021, 1, 1),
catchup=False,
tags=['earnings'],
)
Greenstand's Airflow development environment is set up like this:
There is also a dev and prod / prod readonly PostgreSQL database that is used by Airflow. Some Airflow DAGs transform the data in the PostgreSQL databases.
local Airflow and dev Airflow use the dev PostgreSQL database.
prod Airflow uses the prod PostgreSQL database.
dev Airflow is deployed on Greenstand's dev Kubernetes cluster on DigitalOcean.
prod Airflow is deployed on Greenstand's prod Kubernetes cluster on DigitalOcean.
Please see: https://github.com/Greenstand/treetracker-infrastructure/tree/master/airflow for how Airflow is configured and installed on Kubernetes. Greenstand uses Ansible to deploy and configure an Airflow Helm Chart onto the Kubernetes cluster.
The local Airflow development environment is for Greenstand volunteers to develop Airflow DAGs locally on their machine. Once you have finished developing locally, you can create a Pull Request on Github and merge the changes back to the main branch of this repo. Changes to the main branch will automatically be deployed to the dev Airflow. Once the new DAG is working on dev Airflow, the production branch of this repo will be updated and the new changes will be deployed to the prod Airflow.
For access credentials to the dev Airflow and dev or prod readonly postgreSQL database: