YoVDO

Fine-Tuning Large Language Models with Limited Resources - NeurIPS Hacker Cup AI

Offered By: Weights & Biases via YouTube

Tags

Fine-Tuning Courses LoRA (Low-Rank Adaptation) Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the intricacies of fine-tuning large language models under resource constraints in this live session from the NeurIPS Hacker Cup AI competition. Delve into optimization techniques like gradient accumulation, activation checkpointing, and LoRa as Joe Cunnings from Meta's torchtune team shares strategies for maximizing model performance with limited hardware. Learn the importance of high-quality datasets and gain practical advice for working within a 40 GB VRAM GPU environment. Perfect for developers seeking to enhance their skills in efficient model fine-tuning and competition-ready AI development. Access the torchtune GitHub repository at https://github.com/pytorch/torchtune for additional resources and tools.

Syllabus

NeurIPS Hacker Cup AI: FineTuning


Taught by

Weights & Biases

Related Courses

TensorFlow: Working with NLP
LinkedIn Learning
Introduction to Video Editing - Video Editing Tutorials
Great Learning via YouTube
HuggingFace Crash Course - Sentiment Analysis, Model Hub, Fine Tuning
Python Engineer via YouTube
GPT3 and Finetuning the Core Objective Functions - A Deep Dive
David Shapiro ~ AI via YouTube
How to Build a Q&A AI in Python - Open-Domain Question-Answering
James Briggs via YouTube