RLHF: How to Learn from Human Feedback with Reinforcement Learning
Offered By: Cooperative AI Foundation via YouTube
Course Description
Overview
Explore the intricacies of Reinforcement Learning from Human Feedback (RLHF) in this 59-minute lecture delivered at the 2023 Cooperative AI Summer School. Delve into the innovative techniques for leveraging human input to enhance AI systems as presented by Natasha Jaques, a Senior Research Scientist at Google Brain. Learn about the applications of RLHF in multi-agent and human-AI interactions, drawing from Jaques' extensive research background and accolades in the field of Social Reinforcement Learning. Gain insights from her experiences at prestigious institutions like MIT, UC Berkeley, DeepMind, and OpenAI, and discover how RLHF is shaping the future of AI development and human-machine collaboration.
Syllabus
RLHF: How to Learn from Human Feedback with Reinforcement Learning
Taught by
Cooperative AI Foundation
Related Courses
Direct Preference Optimization - Fine-Tuning LLMs Without Reinforcement LearningSerrano.Academy via YouTube Mastering ChatGPT (AI) and PowerPoint presentation
Udemy Reinforcement Learning with TorchRL and TensoDict - NeurIPS Hacker Cup AI
Weights & Biases via YouTube Reinforcement Learning from Human Feedback (RLHF) Explained
IBM via YouTube PUZZLE: Efficiently Aligning Large Language Models through Light-Weight Context Switching
USENIX via YouTube