Streaming Sequence Transduction through Dynamic Compression
Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube
Course Description
Overview
Explore a groundbreaking Transformer-based model called STAR (Stream Transduction with Anchor Representations) in this 46-minute lecture from the Center for Language & Speech Processing at Johns Hopkins University. Discover how STAR efficiently performs sequence-to-sequence transduction over streams by dynamically segmenting input streams and creating compressed anchor representations. Learn about its impressive performance in Automatic Speech Recognition (ASR), achieving nearly lossless compression at 12x and surpassing existing methods. Examine STAR's superior capabilities in simultaneous speech-to-text tasks, including improved segmentation and optimized latency-quality trade-offs. Gain insights into how this innovative model balances latency, memory footprint, and quality for streaming sequence transduction applications.
Syllabus
Streaming Sequence Transduction through Dynamic Compression
Taught by
Center for Language & Speech Processing(CLSP), JHU
Related Courses
Simple Recurrent Neural Network with KerasCoursera Project Network via Coursera Deep Learning: Advanced Natural Language Processing and RNNs
Udemy Machine Translation
Great Learning via YouTube Tensorflow 2.0 | Recurrent Neural Networks, LSTMs, GRUs
Udemy Pytorch Transformers for Machine Translation
Aladdin Persson via YouTube