Evaluate and Choose the Best LLM Using Automatic Metrics on Custom Datasets
Offered By: Venelin Valkov via YouTube
Course Description
Overview
Discover effective methods for evaluating Large Language Models (LLMs) using automated metrics on custom datasets in this 22-minute tutorial. Explore best practices for selecting the optimal LLM for specific projects and assess their performance across various tasks. Gain insights into different evaluation approaches, available tools, and metrics. Follow along with a hands-on demonstration using Google Colab, covering dataset preparation, model prediction generation, naive evaluation techniques, and leveraging AI for AI evaluation. Conclude with a comprehensive evaluation report to make informed decisions when choosing the best LLM for your needs.
Syllabus
- Intro
- Text tutorial on MLExpert.io
- LLM evaluation approaches
- Available tools & metrics
- Evaluation process
- Google Colab setup
- Dataset
- Generate model predictions
- Naive evaluation
- Use AI to evaluate AI
- Evaluation report
- Conclusion
Taught by
Venelin Valkov
Related Courses
Macroeconometric ForecastingInternational Monetary Fund via edX Machine Learning With Big Data
University of California, San Diego via Coursera Data Science at Scale - Capstone Project
University of Washington via Coursera Structural Equation Model and its Applications | 结构方程模型及其应用 (粤语)
The Chinese University of Hong Kong via Coursera Data Science in Action - Building a Predictive Churn Model
SAP Learning