Host Your Own Llama 3 Chatbot in 10 Minutes with Runpod and vLLM - Lecture 3
Offered By: Data Centric via YouTube
Course Description
Overview
Learn how to host a Llama 3 8B model chatbot in just 20 minutes using vLLM's inference server, Runpod GPUs, and Chainlit for the front end. Discover the process of hosting the Llama 3 model on Runpod and creating an efficient chatbot without relying on heavy frameworks. Follow along as the video guides you through building a Runpod template, deploying it, obtaining the endpoint, preparing the Python script, and finally launching the chatbot. Gain practical insights into AI engineering and model hosting, with additional resources provided for further learning and development.
Syllabus
Intro:
Build Runpod Template:
Deploy Runpod Template:
Getting the Endpoint:
Prepping the Python Script:
Launching the Chatbot:
Taught by
Data Centric
Related Courses
Epic Web UI DreamBooth Update - New Best Settings - Stable Diffusion Training Compared on RunPodsSoftware Engineering Courses - SE Courses via YouTube How to Train Stable Diffusion on Your Photos on a Remote GPU - Using RunPod and Dreambooth
AI Tutorials with Kris Kashtanova via YouTube Train Stable Diffusion on Your Own Photos - Updated Tutorial
AI Tutorials with Kris Kashtanova via YouTube ComfyUI Master Tutorial - Stable Diffusion XL - Install on PC, Google Colab and RunPod
Software Engineering Courses - SE Courses via YouTube Stable Diffusion- Training SDXL 1.0 - Finetune, LoRA, D-Adaptation, Prodigy
kasukanra via YouTube