Lies, Damned Lies, and Large Language Models - Measuring and Reducing Hallucinations
Offered By: PyCon US via YouTube
Course Description
Overview
Explore the challenges and solutions surrounding large language models' (LLMs) tendency to produce incorrect information in this 29-minute PyCon US talk. Discover methods to measure and compare hallucination rates among different models, focusing on misinformation regurgitation from training data. Learn to utilize Python tools like Hugging Face's datasets and transformers packages, as well as the langchain package, to assess hallucinations using the TruthfulQA dataset. Gain insights into recent initiatives aimed at reducing hallucinations, including retrieval augmented generation (RAG), and understand how these techniques can enhance LLM reliability across various applications. Access accompanying slides for a comprehensive overview of the presentation's key points and examples.
Syllabus
Talks - Jodie Burchell: Lies, damned lies and large language models
Taught by
PyCon US
Related Courses
Pinecone Vercel Starter Template and RAG - Live Code Review Part 2Pinecone via YouTube Will LLMs Kill Search? The Future of Information Retrieval
Aleksa Gordić - The AI Epiphany via YouTube RAG But Better: Rerankers with Cohere AI - Improving Retrieval Pipelines
James Briggs via YouTube Advanced RAG - Contextual Compressors and Filters - Lecture 4
Sam Witteveen via YouTube LangChain Multi-Query Retriever for RAG - Advanced Technique for Broader Vector Space Search
James Briggs via YouTube