Global Intelligence Pipeline - Crafting Inference at the Edge
Offered By: Conf42 via YouTube
Course Description
Overview
          Explore the cutting-edge world of edge computing and AI inference in this conference talk from Conf42 ML 2024. Delve into the concept of a Global Intelligence Pipeline, learning how to craft inference solutions at the edge for low-latency, real-time processing. Discover Gcore's edge AI solutions, including the use of NVIDIA H100 and A100 GPUs with InfiniBand. Examine market trends in AI adoption, real-time LLM inference examples, and specific use cases for edge AI. Gain insights into solving challenges in LLM inference, understanding network latency impacts, and implementing real-time end-to-end processing. Witness a demonstration of AIoT architecture and learn about achieving optimal network latency goals for edge inference applications.
        
Syllabus
 intro
 preamble
 gcore at a glance
 gcore edge ai solutions
 global intelligence pipeline
 nvidia h100 and a100 + infiniband gpu
 where can i serve my trained model with low latency?
 market overview: increasing rever adopting ai
 real-time llm inference example
 ai use case at the edge
 edge ai inference requirements
 solving challenges in llm inference
 network latency
 real-time end-to-end processing
 aiot architecture
 demo
 inference at the edge
 network latency goal
 thank you!
Taught by
Conf42
Related Courses
Fog Networks and the Internet of ThingsPrinceton University via Coursera AWS IoT: Developing and Deploying an Internet of Things
Amazon Web Services via edX Business Considerations for 5G with Edge, IoT, and AI
Linux Foundation via edX 5G Strategy for Business Leaders
Linux Foundation via edX Intel® Edge AI Fundamentals with OpenVINO™
Intel via Udacity
