https://github.com/status-im/airflow-dags
Status BI python DAGs for Airflow
https://github.com/status-im/airflow-dags
airflow airflow-dags business-intelligence
Last synced: 6 months ago
JSON representation
Status BI python DAGs for Airflow
- Host: GitHub
- URL: https://github.com/status-im/airflow-dags
- Owner: status-im
- Created: 2021-03-12T09:25:57.000Z (over 4 years ago)
- Default Branch: prod
- Last Pushed: 2025-04-03T13:27:33.000Z (6 months ago)
- Last Synced: 2025-04-05T22:17:18.486Z (6 months ago)
- Topics: airflow, airflow-dags, business-intelligence
- Language: Python
- Homepage:
- Size: 177 KB
- Stars: 0
- Watchers: 9
- Forks: 2
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Description
This repo contains implementations of Airflow workflows and tasks called respectively [DAGs](https://airflow.apache.org/docs/apache-airflow/stable/concepts.html#dags) and [Operators](https://airflow.apache.org/docs/apache-airflow/stable/concepts.html#operators).
* DAGs - Direct Acyclic Graphs - Python scripts defining workflows in a way that reflects their relationships.
* Operators - Python functions which define the individual tasks that are executed as part of a DAG run.To learn how to write DAGs and Operators read about [core concepts](https://airflow.apache.org/docs/apache-airflow/stable/concepts.html#concepts) and follow the [official tutorial](https://airflow.apache.org/docs/apache-airflow/stable/tutorial.html).
# DAG
This repository contains:
* `website_sync`: DAG to launch the Airbyte jobs for the status-website charts.
* `spiff_sync`: DAG to synchronize Spiff workflows data
* `dbt`: DAG to run all the dbt models,
* `gh_sync`: DAG to synchronize data from repository (logos, waku, codex)
* `status_analytics`: DAG to extract and clean data from the google play store and Mixpanel
* `shopify`: DAG to extract and clean data from Shopify
* `budget`: DAG to extract and clean data from an excel sheet where there are budget transactions and forecast data fron Finance team.> The DBT models run in some DAG are stored in [`dbt-models`](https://github.com/status-im/dbt-models).
# Continuous Integration
Changes pushed to `master` are automatically fetched to our Airflow instance by the [`airflow-webhook`](https://github.com/status-im/infra-bi/tree/master/ansible/roles/airflow-webhook) service.
# Branches
This repos has 3 working branches:
* `prod`: used by https://airflow.bi.status.im.
* `test`: used by https://airflow.test.bi.status.im to test DAGs modification.
* `example`: contains examples of DAGs# Infrastructure
All Airflow infrastructure is managed in the [infra-bi](https://github.com/status-im/infra-bi) repository.