Fine-Tuning Large Language Models at Scale - Workday's Approach
Offered By: Anyscale via YouTube
Course Description
Overview
Discover Workday's innovative approach to fine-tuning Large Language Models (LLMs) at scale in this Ray Summit 2024 presentation. Explore how Trevor DiMartino addresses the challenges of training models on isolated customer data within a secure, multi-tenant environment while managing GPU scarcity and strict data access controls. Learn about Workday's platform design, which utilizes Parameter-Efficient Fine-Tuning (PEFT) techniques like LoRA and KubeRay's autoscaling capabilities to enable cost-efficient, on-demand GPU resource allocation for both research and production environments. Gain insights into how Ray is leveraged at various scales to create a flexible deployment solution, making full-stack development as accessible as full-scale production in Workday's multi-tenant ecosystem.
Syllabus
How Workday Fine-Tunes LLMs at Scale | Ray Summit 2024
Taught by
Anyscale
Related Courses
Sailing Ray Workloads with KubeRay and Kueue in KubernetesCNCF [Cloud Native Computing Foundation] via YouTube Accelerate Your GenAI Model Inference with Ray and Kubernetes
CNCF [Cloud Native Computing Foundation] via YouTube Forecasting Covid Infections for the UK's NHS Using Ray and Kubernetes
Anyscale via YouTube KubeRay: A Ray Cluster Management Solution on Kubernetes
Anyscale via YouTube The Different Shades of Using KubeRay with Kubernetes
Anyscale via YouTube