Re-Imagining Apache Spark Development - Tools for Productivity and Standardization
Offered By: Databricks via YouTube
Course Description
Overview
Explore a 25-minute conference talk that challenges traditional ETL tools and proposes a new approach to Apache Spark development. Delve into the evolution of data engineering practices, from ETL tools to code-based solutions, and discover why current methods may be falling short. Learn about innovative tools designed to enhance Spark development, focusing on productivity, code standardization, metadata management, lineage tracking, and agile CI/CD processes. Gain insights into the potential of a new generation of development tools that combine the benefits of code-based approaches with the standardization and productivity features of traditional ETL tools. Witness a demonstration of Prophecy, a tool embodying these new principles, and understand how it aims to revolutionize Apache Spark development for modern data engineering needs.
Syllabus
Introduction
Data Engineering vs ETL
How to become successful with ETL
Its bad for Spark
This is 2020
What does Butdo look like
Engineering Tools
Visual ETL
Standardized Components
Metadata
Continuous Deployment
Compilers
Demo
Prophecy
Taught by
Databricks
Related Courses
ETL مع Python : استخراج وتحويل وتحميل البياناتCoursera Project Network via Coursera AWS: Data Collection Systems
Whizlabs via Coursera AWS Glue Getting Started (Indonesian)
Amazon Web Services via AWS Skill Builder AWS SimuLearn: Application Logs
Amazon Web Services via AWS Skill Builder AWS SimuLearn: Business Intelligence Dashboards
Amazon Web Services via AWS Skill Builder