YoVDO

Lies, Damned Lies, and Large Language Models - Measuring and Reducing Hallucinations

Offered By: PyCon US via YouTube

Tags

Machine Learning Courses Python Courses LangChain Courses Misinformation Courses Hugging Face Courses Retrieval Augmented Generation Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the challenges and solutions surrounding large language models' (LLMs) tendency to produce incorrect information in this 29-minute PyCon US talk. Discover methods to measure and compare hallucination rates among different models, focusing on misinformation regurgitation from training data. Learn to utilize Python tools like Hugging Face's datasets and transformers packages, as well as the langchain package, to assess hallucinations using the TruthfulQA dataset. Gain insights into recent initiatives aimed at reducing hallucinations, including retrieval augmented generation (RAG), and understand how these techniques can enhance LLM reliability across various applications. Access accompanying slides for a comprehensive overview of the presentation's key points and examples.

Syllabus

Talks - Jodie Burchell: Lies, damned lies and large language models


Taught by

PyCon US

Related Courses

Deception Detox - using research methods and statistics to change the world
Erasmus University Rotterdam via Coursera
Fake News, Facts, and Alternative Facts
University of Michigan via Coursera
Disinformation, Misinformation, and Fake News Teach-Out
University of Michigan via Coursera
Disinformation, Misinformation, and Fake News Teach-Out
University of Michigan via FutureLearn
Responsible News Consumption in the Digital Age: Strategies for Citizens of Post-Soviet States
Higher School of Economics via Coursera