The State of vLLM - Advancements in LLM Inference and Serving
Offered By: Anyscale via YouTube
Course Description
Overview
Explore the latest developments in vLLM, the open-source LLM inference and serving engine, in this 35-minute conference talk from Ray Summit 2024. Join Kuntai Du from the University of Chicago and Zhuohan Li from UC Berkeley as they delve into the significant progress made by vLLM over the past year. Discover the project's growing adoption, new features, and performance improvements. Gain insights into the community growth and governance changes shaping vLLM's ecosystem. Learn about the roadmap for upcoming releases and get a glimpse into the future of this rapidly evolving LLM serving solution. Ideal for those interested in efficient LLM deployment and serving technologies, this presentation offers valuable information on the cutting-edge advancements in the field.
Syllabus
The State of vLLM | Ray Summit 2024
Taught by
Anyscale
Related Courses
Cloud Computing Concepts, Part 1University of Illinois at Urbana-Champaign via Coursera Cloud Computing Concepts: Part 2
University of Illinois at Urbana-Champaign via Coursera Reliable Distributed Algorithms - Part 1
KTH Royal Institute of Technology via edX Introduction to Apache Spark and AWS
University of London International Programmes via Coursera Réalisez des calculs distribués sur des données massives
CentraleSupélec via OpenClassrooms