YoVDO

Fine-tuning LLMs to Reduce Hallucination - Leveraging Out-of-Domain Data

Offered By: Weights & Biases via YouTube

Tags

Fine-Tuning Courses Model Evaluation Courses Language Models Courses Wikipedia Courses Mistral AI Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore techniques for reducing hallucinations in large language models through fine-tuning in this hour-long webinar from Weights & Biases. Learn how to leverage out-of-domain data to improve MistralAI models' ability to detect factual inconsistencies. Follow along with a hands-on demonstration that covers creating initial prompts, setting up evaluation pipelines, and implementing a two-step fine-tuning process using the Factual Inconsistency Benchmark dataset and Wikipedia summaries. Discover how Weights & Biases Weave can automate model evaluation and see examples of innovative fine-tuning applications from hackathon winning projects. Gain insights into Mistral's latest models, fine-tuning services, and conversational AI tools to enhance natural language inference in production environments.

Syllabus

Webinar agenda and overview of Mistral AI
Fine-Tuning Services: Introduction to Mistral's fine-tuning API and services
Conversational AI Interface: Introduction to LAT, Mistral's conversational AI tool
Latest Model Releases: Newest Mistral models and their features
Fine-Tuning Process: Steps and benefits of fine-tuning models
Hackathon Winning Projects: Examples of innovative uses of fine-tuning
Hands-On Demo Introduction: Introduction to the practical demo segment
Setting Up the Demo: Instructions for setting up and running the demo notebook
Creating Initial Prompt: Steps to create and test an initial prompt
Evaluation Pipeline: Setting up and running an evaluation pipeline for model performance
Improving Model Performance: Strategies and techniques to enhance model accuracy
Fine-Tuning and Results: Creating and evaluating a fine-tuned model
Two-Step Fine-Tuning: Explanation and demonstration of the two-step fine-tuning process
Conclusion and final thoughts


Taught by

Weights & Biases

Related Courses

Amazon SageMaker JumpStart Foundations (Japanese)
Amazon Web Services via AWS Skill Builder
AWS Flash - Generative AI with Diffusion Models
Amazon Web Services via AWS Skill Builder
AWS Flash - Operationalize Generative AI Applications (FMOps/LLMOps)
Amazon Web Services via AWS Skill Builder
AWS SimuLearn: Automate Fine-Tuning of an LLM
Amazon Web Services via AWS Skill Builder
AWS SimuLearn: Fine-Tune a Base Model with RLHF
Amazon Web Services via AWS Skill Builder