Building Multimodal AI RAG with LlamaIndex, NVIDIA NIM, and Milvus - LLM App Development
Offered By: Nvidia via YouTube
Course Description
Overview
Explore the process of building a multimodal AI retrieval-augmented generation (RAG) application in this 17-minute video tutorial. Learn how to convert documents into text using vision language models like NeVA 22B and DePlot, utilize GPU-accelerated Milvus for efficient embedding storage and retrieval, leverage NVIDIA NIM API's Llama 3 model for handling user queries, and seamlessly integrate all components with LlamaIndex. Gain practical insights into document processing, vector database management, inference techniques, and orchestration for creating a smooth Q&A experience. Access the accompanying notebook for hands-on practice and join the NVIDIA Developer Program for additional resources. Discover how to combine cutting-edge technologies such as LangChain, Mixtral, and NIM APIs to develop advanced LLM applications.
Syllabus
Building Multimodal AI RAG with LlamaIndex, NVIDIA NIM, and Milvus | LLM App Development
Taught by
NVIDIA Developer
Tags
Related Courses
Generative AI, from GANs to CLIP, with Python and PytorchUdemy ODSC East 2022 Keynote by Luis Vargas, Ph.D. - The Big Wave of AI at Scale
Open Data Science via YouTube Comparing AI Image Caption Models: GIT, BLIP, and ViT+GPT2
1littlecoder via YouTube In Conversation with the Godfather of AI
Collision Conference via YouTube LLaVA: The New Open Access Multimodal AI Model
1littlecoder via YouTube