YoVDO

EfficientML.ai: Quantization Part II - Lecture 6

Offered By: MIT HAN Lab via YouTube

Tags

Quantization Courses Machine Learning Courses Neural Networks Courses Convolution Courses Fine-Tuning Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Dive into the second part of a comprehensive lecture on quantization in machine learning, delivered by Prof. Song Han as part of MIT's 6.5940 course for Fall 2023. Explore advanced concepts such as Rooney Quantization, scaling factors, convolution techniques, and post-training quantization methods. Learn about quantization granularity, per-channel quantization, and the importance of clipping in the quantization process. Discover how to select optimal clipping ranges and fine-tune quantized models. Examine weight and activation quantization through practical examples, and gain insights into binary and ternary quantization techniques, including stochastic binarization. Access accompanying slides at efficientml.ai for a deeper understanding of these cutting-edge quantization strategies in efficient machine learning.

Syllabus

Introduction
Outline
Agenda
Rooney Quantization
Original Weight
Scaling Factor
Convolution
Posttraining quantization
Quantization granularity
Perchannel quantization
Skating Factor
Clipping clipping clipping
Selecting clipping range
Fine tuning
Weight
Activation
Quantization Example
Quantization Notation
Quantization Results
Binary ternary quantization
Stochastic binarization


Taught by

MIT HAN Lab

Related Courses

Computational Photography
Georgia Institute of Technology via Udacity
Discrete Time Signals and Systems, Part 1: Time Domain
Rice University via edX
Signals and Systems, Part 1
Indian Institute of Technology Bombay via edX
Discrete Time Signals and Systems, Part 2: Frequency Domain
Rice University via edX
Introduction to Sound and Acoustic Sketching
University St. Joseph via Kadenze