YoVDO

Evaluate and Choose the Best LLM Using Automatic Metrics on Custom Datasets

Offered By: Venelin Valkov via YouTube

Tags

Model Evaluation Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Discover effective methods for evaluating Large Language Models (LLMs) using automated metrics on custom datasets in this 22-minute tutorial. Explore best practices for selecting the optimal LLM for specific projects and assess their performance across various tasks. Gain insights into different evaluation approaches, available tools, and metrics. Follow along with a hands-on demonstration using Google Colab, covering dataset preparation, model prediction generation, naive evaluation techniques, and leveraging AI for AI evaluation. Conclude with a comprehensive evaluation report to make informed decisions when choosing the best LLM for your needs.

Syllabus

- Intro
- Text tutorial on MLExpert.io
- LLM evaluation approaches
- Available tools & metrics
- Evaluation process
- Google Colab setup
- Dataset
- Generate model predictions
- Naive evaluation
- Use AI to evaluate AI
- Evaluation report
- Conclusion


Taught by

Venelin Valkov

Related Courses

Macroeconometric Forecasting
International Monetary Fund via edX
Machine Learning With Big Data
University of California, San Diego via Coursera
Data Science at Scale - Capstone Project
University of Washington via Coursera
Structural Equation Model and its Applications | 结构方程模型及其应用 (粤语)
The Chinese University of Hong Kong via Coursera
Data Science in Action - Building a Predictive Churn Model
SAP Learning