TinyML and Efficient Deep Learning Computing - Course Summary
Offered By: MIT HAN Lab via YouTube
Course Description
Overview
Explore the key concepts and techniques covered in this MIT course on TinyML and Efficient Deep Learning Computing. Gain insights into deploying neural networks on resource-constrained devices like mobile phones and IoT devices. Dive deep into efficient machine learning techniques, including model compression, pruning, quantization, neural architecture search, and distillation. Learn about efficient training methods such as gradient compression and on-device transfer learning. Discover application-specific model optimization for videos, point clouds, and NLP. Get hands-on experience implementing deep learning applications on microcontrollers, mobile phones, and quantum machines through an open-ended design project. Access course materials, including lecture slides, on the course website to enhance your understanding of efficient machine learning techniques for powerful deep learning applications.
Syllabus
Introduction
Efficient Difference Techniques
Summary
Taught by
MIT HAN Lab
Related Courses
Machine Learning Modeling Pipelines in ProductionDeepLearning.AI via Coursera MLOps for Scaling TinyML
Harvard University via edX Parameter Prediction for Unseen Deep Architectures - With First Author Boris Knyazev
Yannic Kilcher via YouTube SpineNet - Learning Scale-Permuted Backbone for Recognition and Localization
Yannic Kilcher via YouTube Synthetic Petri Dish - A Novel Surrogate Model for Rapid Architecture Search
Yannic Kilcher via YouTube