Iterative Preference Learning Methods for Large Language Model Post-Training
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore the intricacies of Reinforcement Learning from Human Feedback (RLHF) in this 53-minute talk by Wei Xiong from UIUC, presented at the Simons Institute. Delve into the mathematical foundations of RLHF, examining its formulation as a reverse-KL regularized contextual bandit problem and its statistical efficiency. Discover how continuous online exploration through human evaluator interactions enhances RLHF's effectiveness. Learn about a novel, provably efficient online iterative training framework that spawns innovative RLHF algorithms like iterative direct preference learning. Gain practical insights into creating state-of-the-art chatbots using open-source data, as demonstrated in the RLHFlow project. This talk, part of the "Emerging Generalization Settings" series, offers a deep dive into the cutting-edge techniques aligning large language models with human preferences.
Syllabus
Iterative preference learning methods for large language model post training
Taught by
Simons Institute
Related Courses
E-learning and Digital CulturesUniversity of Edinburgh via Coursera Construcción de un Curso Virtual en la Plataforma Moodle
Universidad de San Martín de Porres via Miríadax Teaching Computing: Part 2
University of East Anglia via FutureLearn Learning Design
University of Leicester via EMMA Nuevos escenarios de aprendizaje digital
University of the Basque Country via Miríadax