YoVDO

HyperTransformer - Model Generation for Supervised and Semi-Supervised Few-Shot Learning

Offered By: Yannic Kilcher via YouTube

Tags

Few-shot Learning Courses Supervised Learning Courses Semi-supervised Learning Courses

Course Description

Overview

Explore the innovative HyperTransformer model for few-shot learning in this comprehensive video featuring a paper explanation and author interview. Dive into the architecture that generates convolutional neural network weights directly from support samples, decoupling task space complexity from individual task complexity. Learn about weight generation versus fine-tuning, the benefits of self-attention mechanisms, and the potential for producing larger weights. Analyze experimental results, discuss open questions, and gain insights into the model's effectiveness for small target CNN architectures and semi-supervised learning scenarios.

Syllabus

- Intro & Overview
- Weight-generation vs Fine-tuning for few-shot learning
- HyperTransformer model architecture overview
- Why the self-attention mechanism is useful here
- Start of Interview
- Can neural networks even produce weights of other networks?
- How complex does the computational graph get?
- Why are transformers particularly good here?
- What can the attention maps tell us about the algorithm?
- How could we produce larger weights?
- Diving into experimental results
- What questions remain open?


Taught by

Yannic Kilcher

Related Courses

Stanford Seminar - Enabling NLP, Machine Learning, and Few-Shot Learning Using Associative Processing
Stanford University via YouTube
GUI-Based Few Shot Classification Model Trainer - Demo
James Briggs via YouTube
GPT-3 - Language Models Are Few-Shot Learners
Yannic Kilcher via YouTube
IMAML- Meta-Learning with Implicit Gradients
Yannic Kilcher via YouTube
Few-Shot Learning in Production
HuggingFace via YouTube