LLMOps: OpenVino Toolkit para Quantizar LLama 3.2 3B a 4int e Inferencia en CPU
Offered By: The Machine Learning Engineer via YouTube
Course Description
Overview
Explore a 32-minute video tutorial on LLMOps, focusing on using the OpenVino Toolkit to quantize the LLAMA3.2 3B model to 4-bit integer format and perform CPU inference. Learn how to convert the LLAMA3.2 3 billion parameter model to OpenVino IR format, apply 4-bit integer quantization, and execute inference on CPU using Chain of Thought (CoT) prompts. Access the accompanying Jupyter notebook for hands-on practice and in-depth understanding of the process. Ideal for data scientists and machine learning enthusiasts looking to optimize large language models for efficient deployment.
Syllabus
LLMOps: OpenVino Toolkit quantizar 4int LLama3.2 3B e Inferencia CPU #datascience #machinelearning
Taught by
The Machine Learning Engineer
Related Courses
Digital Signal ProcessingÉcole Polytechnique Fédérale de Lausanne via Coursera Principles of Communication Systems - I
Indian Institute of Technology Kanpur via Swayam Digital Signal Processing 2: Filtering
École Polytechnique Fédérale de Lausanne via Coursera Digital Signal Processing 3: Analog vs Digital
École Polytechnique Fédérale de Lausanne via Coursera Digital Signal Processing 4: Applications
École Polytechnique Fédérale de Lausanne via Coursera