YoVDO

The Lottery Ticket Hypothesis for Gigantic Pre-Trained Models

Offered By: VinAI via YouTube

Tags

Pre-trained Models Courses Computer Vision Courses BERT Courses Transfer Learning Courses ImageNet Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the lottery ticket hypothesis for large pre-trained models in this seminar presented by Professor Atlas Wang from UT Austin. Delve into the fascinating world of machine learning, computer vision, and optimization as Wang discusses his research on finding smaller, trainable subnetworks within enormous pre-trained models. Learn about the application of this hypothesis to both NLP and computer vision domains, with specific examples from BERT and ImageNet pre-trained models. Discover how these subnetworks can achieve high levels of sparsity while maintaining full accuracy and transferability to downstream tasks. Gain insights into the implications of this research for the future of deep learning and large-scale pre-training paradigms. Access additional resources and project details through the provided webpage link.

Syllabus

Seminar Series: The lottery ticket hypothesis for gigantic pre trained models


Taught by

VinAI

Related Courses

Inference with Torch-TensorRT Deep Learning Prediction for Beginners - CPU vs CUDA vs TensorRT
Python Simplified via YouTube
AlexNet and ImageNet Explained
James Briggs via YouTube
Analysis of Large-Scale Visual Recognition - Bay Area Vision Meeting
Meta via YouTube
Introduction to Neural Networks for Computer Vision - Part I
University of Central Florida via YouTube
Evaluating Neural Network Robustness - Targeted Attacks and Defenses
University of Central Florida via YouTube