Airflow + dbt + Snowflake Data Pipeline
Production-ready end-to-end data pipeline orchestration using Apache Airflow with Astronomer Cosmos. Implements dbt for analytics engineering and data transformation in Snowflake cloud data warehouse.
Features include dimensional modeling with staging layers and fact/dimension tables, automated data quality tests, incremental loads, GitOps workflow with containerized deployments, and comprehensive monitoring with custom logging.
The project uses Astronomer's Astro Runtime for Airflow deployment in Docker containers, with a dedicated dbt virtual environment and automated dependency management. DAG scheduling runs daily with intelligent task orchestration.
Key Features
- Automated ELT pipeline with dbt transformations orchestrated by Airflow
- Dimensional data modeling with staging, intermediate, and mart layers
- Comprehensive data quality testing with custom SQL tests and generic schema tests
- Containerized deployment using Docker with Astronomer Astro Runtime
- Modular project structure with reusable macros and dbt packages (dbt_utils)
- Snowflake integration with optimized warehouse utilization and role-based access