YoVDO

Methods for Evaluating Your GenAI Application Quality

Offered By: Databricks via YouTube

Tags

Generative AI Courses Databricks Courses Quality Assurance Courses LLM (Large Language Model) Courses Model Evaluation Courses MLFlow Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore comprehensive methods for evaluating Generative AI application quality in this 37-minute conference talk by Databricks. Dive into the suite of tools including inference tables, Lakehouse Monitoring, and MLflow for rigorous evaluation and quality assurance of model responses. Learn to conduct offline evaluations and real-time monitoring, ensuring high-performance standards. Discover best practices for using LLMs as judges, integrating MLflow for experiment tracking, and leveraging inference tables and Lilac for enhanced model management. Optimize workflows and ensure robust, scalable GenAI applications aligned with production goals. Presented by Alkis Polyzotis and Michael Carbin, this talk offers valuable insights for developers and data scientists working with Generative AI technologies.

Syllabus

Methods for Evaluating Your GenAI Application Quality


Taught by

Databricks

Related Courses

Web Engineering III: Quality Assurance
Technische Hochschule Mittelhessen via iversity
Project Management for Designers
Emily Carr University of Art + Design via Kadenze
Develop Project Management Skills
OpenLearning
Study UK: A Guide for Education Agents
FutureLearn
Basics of Design Testing
SAP Learning