Version Control for Lakehouse Architecture - Essential Practices and Benefits
Offered By: Databricks via YouTube
Course Description
Overview
Discover how to implement engineering best practices for data products using data version control with lakeFS in this 15-minute conference talk sponsored by lakeFS. Learn why version control is essential for your lakehouse architecture when developing and maintaining data/ML pipelines using Databricks. Explore techniques to improve data quality and velocity, including experimenting during development, testing data quality in isolation, automating quality validation tests, and achieving full reproducibility of data pipelines. Understand how poor data quality or lack of reproducibility can impact products relying on analytics or machine learning. Gain insights from Oz Katz, CTO & Co-creator of lakeFS, on implementing data version control to enhance your data products. Additional resources on the Rise of the Data Lakehouse and Lakehouse Fundamentals Training are provided for further exploration.
Syllabus
Sponsored by: lakeFS | Why Version Control is Essential for Your Lakehouse Architecture
Taught by
Databricks
Related Courses
内存数据库管理openHPI CS115x: Advanced Apache Spark for Data Science and Data Engineering
University of California, Berkeley via edX Processing Big Data with Azure Data Lake Analytics
Microsoft via edX Google Cloud Big Data and Machine Learning Fundamentals en Español
Google Cloud via Coursera Google Cloud Big Data and Machine Learning Fundamentals 日本語版
Google Cloud via Coursera