Evaluating Language Models - Challenges and Best Practices
Offered By: MLOps.community via YouTube
Course Description
Overview
Explore the challenges and solutions for evaluating language models in this 23-minute lightning talk from the AI in Production Conference. Delve into the metrics and datasets available for assessment, and examine the difficulties of continuous evaluation in production environments. Learn about common pitfalls to avoid and gain insights from Matthew Sharp, author of "LLMs in Production" and a seasoned professional with over a decade of experience in ML/AI and deploying models to production. Discover the importance of contributing to public evaluation datasets and join the call for a community-wide effort to reduce harmful bias in language models. Gain valuable takeaways for improving language model evaluation practices in your own projects or organizations.
Syllabus
Evaluating Language Models // Matthew Sharp // AI in Production Conference Lightning Talk
Taught by
MLOps.community
Related Courses
Machine Learning Operations (MLOps): Getting StartedGoogle Cloud via Coursera Проектирование и реализация систем машинного обучения
Higher School of Economics via Coursera Demystifying Machine Learning Operations (MLOps)
Pluralsight Machine Learning Engineer with Microsoft Azure
Microsoft via Udacity Machine Learning Engineering for Production (MLOps)
DeepLearning.AI via Coursera