Machine Learning with Spark on Google Cloud Dataproc
Offered By: Google via Google Cloud Skills Boost
Course Description
Overview
In this lab you will learn how to implement logistic regression using a machine learning library for Apache Spark running on a Google Cloud Dataproc cluster to develop a model for data from a multivariable dataset
Syllabus
- GSP271
- Overview
- Setup and requirements
- Task 1. Create a Dataproc cluster
- Task 2. Set up bucket and start pyspark session
- Task 3. Read and clean up dataset
- Task 4. Develop a logistic regression model
- Task 5. Save and restore a logistic regression model
- Task 6. Predict with the logistic regression model
- Task 7. Examine model behavior
- Task 8. Evaluate the model
- Congratulations!
Tags
Related Courses
Data Wrangling with MongoDBMongoDB via Udacity Getting and Cleaning Data
Johns Hopkins University via Coursera 软件包在流行病学研究中的应用 Using software apps in epidemiological research
Peking University via Coursera Creating an Analytical Dataset
Udacity Implementing ETL with SQL Server Integration Services
Microsoft via edX