Language Models as Zero-Shot Planners - Extracting Actionable Knowledge for Embodied Agents
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore a comprehensive video lecture and interview on using large language models as zero-shot planners for embodied agents. Delve into the VirtualHome environment and learn how to translate unstructured language model outputs into structured grammar for interactive environments. Discover techniques for decomposing high-level tasks into actionable steps without additional training. Examine the challenges of plan evaluation and execution, and understand the contributions of this research. Gain insights from the interview with first author Wenlong Huang, covering topics such as model size impact, output refinement, and the effectiveness of Codex. Analyze experimental results and consider future implications for extracting actionable knowledge from language models in embodied AI applications.
Syllabus
- Intro & Overview
- The VirtualHome environment
- The problem of plan evaluation
- Contributions of this paper
- Start of interview
- How to use language models with environments?
- What does model size matter?
- How to fix the large models' outputs?
- Possible improvements to the translation procedure
- Why does Codex perform so well?
- Diving into experimental results
- Future outlook
Taught by
Yannic Kilcher
Related Courses
Fast Zero Shot Object Detection with OpenAI CLIPJames Briggs via YouTube Tensor Programs - Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
Aleksa Gordić - The AI Epiphany via YouTube DALL-E - Zero-Shot Text-to-Image Generation - Paper Explained
Aleksa Gordić - The AI Epiphany via YouTube Tuning GPT-3 on a Single GPU via Zero-Shot Hyperparameter Transfer
Massachusetts Institute of Technology via YouTube