YoVDO

End-to-End Data Engineering with Apache Airflow, Docker, and Spark Clusters - Using Python, Scala, and Java

Offered By: CodeWithYu via YouTube

Tags

Apache Airflow Courses Python Courses Java Courses Scala Courses Docker Courses Apache Spark Courses Data Engineering Courses Cluster Computing Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Learn to set up and utilize Apache Airflow and Spark Clusters on Docker in this comprehensive video tutorial. Create an end-to-end data engineering project combining Apache Airflow, Docker, Spark Clusters, Scala, Python, and Java. Develop basic jobs using multiple programming languages, submit them to the Spark cluster for processing, and observe live results. Follow along as the instructor guides you through creating Spark jobs with Python, Scala, and Java, as well as building and compiling Scala and Java jobs. Gain hands-on experience in cluster computation and workflow automation, essential skills for big data analytics and data engineering projects.

Syllabus

Introduction
Creating The Spark Cluster and Airflow on Docker
Creating Spark Job with Python
Creating Spark Job with Scala
Building and Compiling Scala Jobs
Creating Spark Job with Java
Building and Compiling Java Jobs
Cluster computation results


Taught by

CodeWithYu

Related Courses

AWS ML Engineer Associate 1.3 Validate Data and Prepare for Modeling (Simplified Chinese)
Amazon Web Services via AWS Skill Builder
Big Data Analysis with Scala and Spark
École Polytechnique Fédérale de Lausanne via Coursera
Build a Question-answering Bot using Generative AI (Indonesian)
Amazon Web Services via AWS Skill Builder
Data Science with Scala
Cognitive Class
Machine learning with Apache SystemML
Cognitive Class