YoVDO

Streaming Sequence Transduction through Dynamic Compression

Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube

Tags

Transformers Courses Sequence to Sequence Models Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a groundbreaking Transformer-based model called STAR (Stream Transduction with Anchor Representations) in this 46-minute lecture from the Center for Language & Speech Processing at Johns Hopkins University. Discover how STAR efficiently performs sequence-to-sequence transduction over streams by dynamically segmenting input streams and creating compressed anchor representations. Learn about its impressive performance in Automatic Speech Recognition (ASR), achieving nearly lossless compression at 12x and surpassing existing methods. Examine STAR's superior capabilities in simultaneous speech-to-text tasks, including improved segmentation and optimized latency-quality trade-offs. Gain insights into how this innovative model balances latency, memory footprint, and quality for streaming sequence transduction applications.

Syllabus

Streaming Sequence Transduction through Dynamic Compression


Taught by

Center for Language & Speech Processing(CLSP), JHU

Related Courses

Attention Mechanism - Deutsch
Google Cloud via Coursera
Gen AI Foundational Models for NLP & Language Understanding
IBM via Coursera
Introduction to Attention-Based Neural Networks
LinkedIn Learning
Encoder-Decoder Architecture
Pluralsight
Building and Deploying Keras Models in a Multi-cloud Environment
Pluralsight