IDEFICS 2 API Endpoint, vLLM vs TGI, and General Fine-tuning Tips
Offered By: Trelis Research via YouTube
Course Description
Overview
Explore the latest developments in AI model deployment, fine-tuning techniques, and transformer architectures in this comprehensive video tutorial. Dive into deploying the IDEFICS multimodal model on TGI for inference, understanding the nuances of prompt engineering for optimal performance. Learn about representation fine-tuning, Lora, and QLoRA methods for model optimization. Gain insights into transformer architectures, building models from scratch, and working with quantized versions for efficient inference. Discover the potential of ORPO (preference fine-tuning + supervised fine-tuning) and explore strategies for training models using various datasets. Benefit from expert answers to common questions and get a preview of upcoming topics in the field of AI research and development.
Syllabus
Introduction
Latest video on representation fine-tuning and fine-tuning with Lora
Deploying IDEFICS multimodal model on TGI for inference
Answering questions from the chat
Transformer architectures and building models from scratch
Quantized model versions and inference
Lora vs QLoRA for fine-tuning
ORPO preference fine-tuning + supervised fine-tuning
Training models from scratch and datasets
Wrap-up and upcoming videos
Taught by
Trelis Research
Related Courses
Fine-Tuning LLM with QLoRA on Single GPU - Training Falcon-7b on ChatBot Support FAQ DatasetVenelin Valkov via YouTube Deploy LLM to Production on Single GPU - REST API for Falcon 7B with QLoRA on Inference Endpoints
Venelin Valkov via YouTube Building an LLM Fine-Tuning Dataset - From Reddit Comments to QLoRA Training
sentdex via YouTube Generative AI: Fine-Tuning LLM Models Crash Course
Krish Naik via YouTube Aligning Open Language Models - Stanford CS25 Lecture
Stanford University via YouTube