YoVDO

Introduction to Optimization Algorithms to Compress Neural Networks

Offered By: tinyML via YouTube

Tags

Deep Learning Courses Embedded Systems Courses Quantization Courses Optimization Algorithms Courses

Course Description

Overview

Explore optimization algorithms for compressing neural networks in this tinyML Talks webcast. Dive into the challenges of deploying advanced networks on resource-constrained systems and learn about various compression techniques. Discover the functionality of common compression algorithms, including pruning, quantization, and knowledge distillation. Examine the pros and cons of different pruning techniques, and explore concepts such as lowend factorization, fast convolutional networks, and selective attention networks. Gain insights into general use cases and the process of pruning whole channels. This comprehensive talk, presented by Marcus Rüb from the Hahn-Schickard Research Institute, provides a valuable introduction to making neural networks more efficient for embedded devices and mobile applications.

Syllabus

Introduction
Local Meetup
What is tinyML
Quantization
Knowledge distillation
Pruning
Pruning techniques
Pros and cons
Lowend factorization
Fast convolutional
Selective attention network
Summary
Questions
General use cases
Pruning whole channels
Conclusion


Taught by

tinyML

Related Courses

Deep Learning for Natural Language Processing
University of Oxford via Independent
Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and Optimization
DeepLearning.AI via Coursera
Deep Learning Part 1 (IITM)
Indian Institute of Technology Madras via Swayam
Deep Learning - Part 1
Indian Institute of Technology, Ropar via Swayam
Logistic Regression with Python and Numpy
Coursera Project Network via Coursera