Apache Airflow Training
Level
IntermediateDuration
16h / 2 daysDate
Individually arrangedPrice
Individually arrangedApache Airflow Training
The Apache Airflow training is an intensive practical course that introduces participants to orchestrating ETL processes and automating workflows in Big Data environments. Participants will learn how to install, configure, and efficiently use Apache Airflow to schedule and monitor tasks in recurring and one-time business processes. The training combines theory with practice, enabling participants to gain the skills needed to effectively manage complex data workflows.
What You Will Learn
- Installation and configuration of Apache Airflow
- Understanding Airflow architecture and components
- Creating DAGs (Directed Acyclic Graphs) and defining tasks
- Passing data between tasks using XCom
- Implementing branching and task organization
- Managing subDAGs and task groups
- Monitoring and debugging workflows
- DAG restart options, configuration resets, and log analysis
Who is this training for?
Data scientists working with large datasets
Developers with Python experience
DevOps specialists responsible for data infrastructure
Data engineers and analysts working on process automation
Training Program
-
Module 1 – Introduction
- History of Apache Airflow (Airflow 1.0 vs Airflow 2.0)
- Overview of Airflow architecture and components
- Core elements: DAG, Instance, Task
-
Module 2 – Creating DAGs
- Operators, Sensors, Hooks
- Creating connections and configuring tasks
-
Module 3 – Advanced Techniques in Airflow
- XCom – passing data between tasks
- Branching – conditional task execution
- Task organization – subDAGs and task groups
- Monitoring and analyzing logs
- DAG restart options and configuration resets
-
Optional Module – Advanced Configuration
- Building and configuring Apache Airflow clusters
- Using advanced DAG features