Speech and Audio Processing in Non-Invasive Brain-Computer Interfaces at Meta
Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube
Course Description
Overview
Explore the potential of non-invasive neural interfaces in transforming human-computer interaction through this 43-minute talk by Michael Mandel from Reality Labs at Meta. Delve into the development of an interface for controlling augmented reality devices using electromyographic (EMG) signals captured at the wrist. Discover how speech and audio technologies are uniquely suited to unlocking the full potential of these signals and interactions. Learn about the neuroscientific background necessary to understand these signals, and examine automatic speech recognition-inspired interfaces for generating text and beamforming-inspired interfaces for identifying individual neurons. Gain insights into how these technologies connect with egocentric machine intelligence tasks that could be implemented on augmented reality devices. Understand the potential for creating effortless and joyful interfaces that provide low friction, information-rich, and always available inputs for users.
Syllabus
Speech and Audio Processing in Non-Invasive Brain-Computer Interfaces at Meta [Michael Mandel]
Taught by
Center for Language & Speech Processing(CLSP), JHU
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Natural Language Processing
Columbia University via Coursera Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent