Version Control for Lakehouse Architecture - Essential Practices and Benefits
Offered By: Databricks via YouTube
Course Description
Overview
Discover how to implement engineering best practices for data products using data version control with lakeFS in this 15-minute conference talk sponsored by lakeFS. Learn why version control is essential for your lakehouse architecture when developing and maintaining data/ML pipelines using Databricks. Explore techniques to improve data quality and velocity, including experimenting during development, testing data quality in isolation, automating quality validation tests, and achieving full reproducibility of data pipelines. Understand how poor data quality or lack of reproducibility can impact products relying on analytics or machine learning. Gain insights from Oz Katz, CTO & Co-creator of lakeFS, on implementing data version control to enhance your data products. Additional resources on the Rise of the Data Lakehouse and Lakehouse Fundamentals Training are provided for further exploration.
Syllabus
Sponsored by: lakeFS | Why Version Control is Essential for Your Lakehouse Architecture
Taught by
Databricks
Related Courses
Data Processing with AzureLearnQuest via Coursera Mejores prácticas para el procesamiento de datos en Big Data
Coursera Project Network via Coursera Data Science with Databricks for Data Analysts
Databricks via Coursera Azure Data Engineer con Databricks y Azure Data Factory
Coursera Project Network via Coursera Curso Completo de Spark con Databricks (Big Data)
Coursera Project Network via Coursera