YoVDO

The Complete Hands-On Introduction to Apache Airflow

Offered By: Udemy

Tags

Apache Airflow Courses Docker Courses Data Pipelines Courses

Course Description

Overview

Learn to author, schedule and monitor data pipelines through practical examples using Apache Airflow

What you'll learn:
  • Create plugins to add functionalities to Apache Airflow.
  • Using Docker with Airflow and different executors
  • Master core functionalities such as DAGs, Operators, Tasks, Workflows, etc
  • Understand and apply advanced concepts of Apache Airflow such as XCOMs, Branching and SubDAGs.
  • The difference between Sequential, Local and Celery Executors, how do they work and how can you use them.
  • Use Apache Airflow in a Big Data ecosystem with Hive, PostgreSQL, Elasticsearch etc.
  • Install and configure Apache Airflow
  • Think, answer and implement solutions using Airflow to real data processing problems

Apache Airflow is an open-source platform to programmatically author, schedule and monitor workflows. If you have many ETL(s) to manage, Airflow is a must-have.

In this course you are going to learn everything you need to start using Apache Airflow through theory and pratical videos. Starting from very basic notions such as, what is Airflow and how it works, we will dive into advanced concepts such as, how to create plugins and make real dynamic pipelines.


Taught by

Marc Lamberti

Related Courses

Building ETL and Data Pipelines with Bash, Airflow and Kafka
IBM via edX
Building Data Engineering Pipelines in Python
DataCamp
Introduction to Airflow in Python
DataCamp
ETL and Data Pipelines with Shell, Airflow and Kafka
IBM via Coursera
Cloud Composer: Copying BigQuery Tables Across Different Locations
Google Cloud via Coursera