YoVDO

The Era of 1-bit LLMs Explained - BitNet b1.58 and New Scaling Laws

Offered By: Unify via YouTube

Tags

Model Compression Courses Neural Networks Courses Quantization Courses Scaling Laws Courses Transformer Architecture Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the groundbreaking research presented in the paper "The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits" during this 58-minute session. Delve into the innovative BitNet b1.58 model, which uses ternary parameters {-1, 0, 1} to match full-precision Transformer LLMs in performance while offering significant cost-effectiveness in latency, memory, throughput, and energy consumption. Discover how this 1.58-bit LLM establishes a new scaling law and training recipe for high-performance, cost-effective large language models. Gain insights from the research led by Shuming Ma and Hongyu Wang at Microsoft, and understand its potential impact on the future of AI development. Learn about additional resources for staying updated on AI research, industry trends, and deployment strategies.

Syllabus

The Era of 1-bit LLMs Explained


Taught by

Unify

Related Courses

Introduction To Mechanical Micro Machining
Indian Institute of Technology, Kharagpur via Swayam
Biomaterials - Intro to Biomedical Engineering
Udemy
OpenAI Whisper - Robust Speech Recognition via Large-Scale Weak Supervision
Aleksa Gordić - The AI Epiphany via YouTube
Turbulence as Gibbs Statistics of Vortex Sheets - Alexander Migdal
Institute for Advanced Study via YouTube
City Analytics - Professor Peter Grindrod CBE
Alan Turing Institute via YouTube