Building Deep Learning Models on Databricks
Offered By: Pluralsight
Course Description
Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
In this course, you will learn to train neural network models using TensorFlow and PyTorch, perform distributed training using the Horovod framework, and perform hyperparameter tuning using Hyperopt.
The Databricks Data Lakehouse platform offers a managed environment to train and compare your deep learning models, perform hyperparameter tuning, and productionize and serve your models. In this course, Building Deep Learning Models on Databricks, you will learn to use Bamboolib for no-code data analysis and transformations. First, you will build deep learning models using TensorFlow 2.0 and Keras, and will create a workspace experiment to manage your runs and use autologging to track model parameters, metrics, and artifacts. Next, you will compare multiple runs to find the best-performing model using the MLflow UI. Then, you will see that in order to have support for autologging in MLflow you need to use the PyTorch Lightning framework to design and train your model. You will also register your model with the model registry and use it for batch inference, deploy a Classic MLflow endpoint to serve model predictions, and use the Horovod framework for distributed training of your model. Finally, you will learn how you can use the Hyperopt tool for hyperparameter tuning of your deep learning models, and will run hyperparameter tuning in a distributed fashion on a Spark cluster using the SparkTrials class. When you are finished with this course, you will have the skills and knowledge to build and train deep machine learning models on Databricks using MLflow to manage your machine learning workflow.
The Databricks Data Lakehouse platform offers a managed environment to train and compare your deep learning models, perform hyperparameter tuning, and productionize and serve your models. In this course, Building Deep Learning Models on Databricks, you will learn to use Bamboolib for no-code data analysis and transformations. First, you will build deep learning models using TensorFlow 2.0 and Keras, and will create a workspace experiment to manage your runs and use autologging to track model parameters, metrics, and artifacts. Next, you will compare multiple runs to find the best-performing model using the MLflow UI. Then, you will see that in order to have support for autologging in MLflow you need to use the PyTorch Lightning framework to design and train your model. You will also register your model with the model registry and use it for batch inference, deploy a Classic MLflow endpoint to serve model predictions, and use the Horovod framework for distributed training of your model. Finally, you will learn how you can use the Hyperopt tool for hyperparameter tuning of your deep learning models, and will run hyperparameter tuning in a distributed fashion on a Spark cluster using the SparkTrials class. When you are finished with this course, you will have the skills and knowledge to build and train deep machine learning models on Databricks using MLflow to manage your machine learning workflow.
Syllabus
- Course Overview 2mins
- Introducing MLflow on Databricks 16mins
- Implementing Deep Learning Models Using TensorFlow and Keras 62mins
- Implementing Deep Learning Models Using PyTorch 45mins
- Hyperparameter Tuning Using Hyperopt 21mins
Taught by
Janani Ravi
Related Courses
DP-100 Part 2 - ModelingA Cloud Guru Aerial Image Segmentation with PyTorch
Coursera Project Network via Coursera AI Capstone Project with Deep Learning
IBM via Coursera Applied Machine Learning
Johns Hopkins University via Coursera Apply Generative Adversarial Networks (GANs)
DeepLearning.AI via Coursera