Faster and Cheaper Offline Batch Inference with Ray
Offered By: Anyscale via YouTube
Course Description
Overview
Explore the capabilities of Ray Data, an open-source library for large-scale data processing in machine learning applications, with a focus on offline batch inference. Learn how to efficiently process terabytes of data using pretrained models, understand the limitations of traditional data processing tools for modern deep learning applications, and discover how to leverage Ray Data for faster and more cost-effective inference compared to solutions like Spark or SageMaker. Delve into the importance of offline inference for LLM workloads and witness a demonstration of an end-to-end offline batch inference use case. Gain insights into user success stories and understand why Ray Data is considered the optimal solution for offline batch inference and processing, particularly when working with unstructured data and deep learning models.
Syllabus
Faster and Cheaper Offline Batch Inference with Ray
Taught by
Anyscale
Related Courses
Optimizing LLM Inference with AWS Trainium, Ray, vLLM, and AnyscaleAnyscale via YouTube Scalable and Cost-Efficient AI Workloads with AWS and Anyscale
Anyscale via YouTube End-to-End LLM Workflows with Anyscale
Anyscale via YouTube Developing and Serving RAG-Based LLM Applications in Production
Anyscale via YouTube Deploying Many Models Efficiently with Ray Serve
Anyscale via YouTube