LLMOPs: Multimodal Prompting and Inference with Phi-3 Vision 128K Instruct on CPU - ONNX 4-Bit Quantization in C#
Offered By: The Machine Learning Engineer via YouTube
Course Description
Overview
Explore multimodal prompting and inference on CPU using Phi 3 Vision 128K Instruct model quantized to 4 bits in ONNX format with C# in this 23-minute video tutorial. Learn how to implement LLMOPs (Large Language Model Operations) for data science and machine learning applications. Access the accompanying code on GitHub to follow along and practice the demonstrated techniques. Gain insights into optimizing inference for resource-constrained environments and leveraging advanced language models for vision-based tasks.
Syllabus
LLMOPs: Inference en CPU Phi3 Vision 128k Intruct ONNX 4bits in C# #datascience #machinelearning
Taught by
The Machine Learning Engineer
Related Courses
Fine-tuning Phi-3 for LeetCode: Dataset Generation and Unsloth ImplementationAll About AI via YouTube LLM News: GPT-4, Project Astra, Veo, Copilot+ PCs, Gemini 1.5 Flash, and Chameleon
Elvis Saravia via YouTube LLM Tool Use - GPT4o-mini, Groq, and Llama.cpp
Trelis Research via YouTube LoRA Fine-tuning Explained - Choosing Parameters and Optimizations
Trelis Research via YouTube Comparing LLAMA 3, Phi 3, and GPT-3.5 Turbo AI Agents for Web Search Performance
Data Centric via YouTube