Speech and Audio Processing in Non-Invasive Brain-Computer Interfaces at Meta
Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube
Course Description
Overview
Explore the potential of non-invasive neural interfaces in transforming human-computer interaction through this 43-minute talk by Michael Mandel from Reality Labs at Meta. Delve into the development of an interface for controlling augmented reality devices using electromyographic (EMG) signals captured at the wrist. Discover how speech and audio technologies are uniquely suited to unlocking the full potential of these signals and interactions. Learn about the neuroscientific background necessary to understand these signals, and examine automatic speech recognition-inspired interfaces for generating text and beamforming-inspired interfaces for identifying individual neurons. Gain insights into how these technologies connect with egocentric machine intelligence tasks that could be implemented on augmented reality devices. Understand the potential for creating effortless and joyful interfaces that provide low friction, information-rich, and always available inputs for users.
Syllabus
Speech and Audio Processing in Non-Invasive Brain-Computer Interfaces at Meta [Michael Mandel]
Taught by
Center for Language & Speech Processing(CLSP), JHU
Related Courses
واجهة الدماغ والحاسوبRwaq (رواق) Introductory Neuroscience & Neuro-Instrumentation
Indian Institute of Science Bangalore via Swayam Brain Computer Interface With Python and OpenBCI
YouTube Brain Computer App with React Native - Live Coding
freeCodeCamp Neural Science for Engineers
NPTEL via Swayam