Image Annotation with LLaVA and Ollama
Offered By: Sam Witteveen via YouTube
Course Description
Overview
Explore image annotation techniques using LLava and Ollama in this 15-minute tutorial video. Learn the basic concept of image captioning and follow a step-by-step process to build an image annotation application. Discover how to retrieve file lists from folders, load files, utilize LLaVA 1.6 via Ollama for image processing, save results to a DataFrame, and export data to CSV. Gain practical insights through code demonstrations and access provided GitHub repositories for further exploration. Perfect for developers interested in implementing LLM agents and enhancing their skills in image processing and annotation.
Syllabus
Intro
Image Captioning
Basic Idea of the Image Captioning ap
Image Captioning Diagram
Step 1: Get the file list from a folder
Step 2: Loading the files
Step 3: Send the file to LLaVA 1.6 via Ollama
Step 4: Saving the results back tothe DataFrame
Step 5: Save the DataFrame to CSV
Code Time
Taught by
Sam Witteveen
Related Courses
LLaVA: The New Open Access Multimodal AI Model1littlecoder via YouTube Autogen and Local LLMs Create Realistic Stable Diffusion Model Autonomously
kasukanra via YouTube Unraveling Multimodality with Large Language Models
Linux Foundation via YouTube Efficient and Portable AI/LLM Inference on the Edge Cloud - Workshop
Linux Foundation via YouTube Training and Serving Custom Multi-modal Models - IDEFICS 2 and LLaVA Llama 3
Trelis Research via YouTube