YoVDO

From Python to PySpark and Back Again - Unifying Single-Host and Distributed Deep Learning with Maggy

Offered By: Databricks via YouTube

Tags

PySpark Courses Deep Learning Courses Python Courses Apache Spark Courses Distributed Systems Courses Software Engineering Courses Hyperparameter Tuning Courses

Course Description

Overview

Explore distributed deep learning with Maggy, an open-source framework that bridges the gap between single-host Python and cluster-scale PySpark programs. Learn how to create reusable training functions that seamlessly transition from laptop development to cluster environments. Discover best practices for writing TensorFlow programs, factoring out dependencies, and implementing popular programming idioms. Experience the power of iterative model development in a single Jupyter notebook, mixing vanilla Python and PySpark-specific code. Gain insights into the benefits of distributed deep learning, including faster training with multiple GPUs, parallel hyperparameter tuning, and ablation studies. Understand how Maggy enables DRY (Don't Repeat Yourself) code principles in training functions, allowing for efficient development across different computing environments. Witness a practical demonstration of Maggy's capabilities and learn how to leverage this framework to streamline your deep learning workflows.

Syllabus

Intro
Model Development
Software Engineering
Transparent Code
Building Blocks
Code
Demonstration
Summary


Taught by

Databricks

Related Courses

CS115x: Advanced Apache Spark for Data Science and Data Engineering
University of California, Berkeley via edX
Big Data Analytics
University of Adelaide via edX
Big Data Essentials: HDFS, MapReduce and Spark RDD
Yandex via Coursera
Big Data Analysis: Hive, Spark SQL, DataFrames and GraphFrames
Yandex via Coursera
Introduction to Apache Spark and AWS
University of London International Programmes via Coursera