Scaling Probabilistically Safe Learning to Robotics
Offered By: Paul G. Allen School via YouTube
Course Description
Overview
Explore cutting-edge developments in scaling safe learning to robotics in this 57-minute Winter 2021 Robotics Colloquium talk by Scott Niekum from the University of Texas, Austin. Delve into three key areas: a theory of safe imitation learning, scalable reward inference without models, and efficient policy evaluation. Discover how these algorithms blend safety and practicality to advance high-confidence robot learning with limited real-world data. Learn about the challenges of deploying learning robots in real-world scenarios and the importance of probabilistic guarantees for safety and performance. Gain insights from Niekum's expertise in imitation learning, reinforcement learning, and robotic manipulation as he discusses the progress towards making safe robot learning more practical and scalable.
Syllabus
Scaling Probabilistically Safe Learning to Robotics (Scott Niekum, University of Texas, Austin)
Taught by
Paul G. Allen School
Related Courses
Computational NeuroscienceUniversity of Washington via Coursera Reinforcement Learning
Brown University via Udacity Reinforcement Learning
Indian Institute of Technology Madras via Swayam FA17: Machine Learning
Georgia Institute of Technology via edX Introduction to Reinforcement Learning
Higher School of Economics via Coursera