YoVDO

Iterative Preference Learning Methods for Large Language Model Post-Training

Offered By: Simons Institute via YouTube

Tags

Online Learning Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the intricacies of Reinforcement Learning from Human Feedback (RLHF) in this 53-minute talk by Wei Xiong from UIUC, presented at the Simons Institute. Delve into the mathematical foundations of RLHF, examining its formulation as a reverse-KL regularized contextual bandit problem and its statistical efficiency. Discover how continuous online exploration through human evaluator interactions enhances RLHF's effectiveness. Learn about a novel, provably efficient online iterative training framework that spawns innovative RLHF algorithms like iterative direct preference learning. Gain practical insights into creating state-of-the-art chatbots using open-source data, as demonstrated in the RLHFlow project. This talk, part of the "Emerging Generalization Settings" series, offers a deep dive into the cutting-edge techniques aligning large language models with human preferences.

Syllabus

Iterative preference learning methods for large language model post training


Taught by

Simons Institute

Related Courses

E-learning and Digital Cultures
University of Edinburgh via Coursera
Construcción de un Curso Virtual en la Plataforma Moodle
Universidad de San Martín de Porres via Miríadax
Teaching Computing: Part 2
University of East Anglia via FutureLearn
Learning Design
University of Leicester via EMMA
Nuevos escenarios de aprendizaje digital
University of the Basque Country via Miríadax