Columbia engineering researchers make progress toward development of cognitively-controlled hearing aids
The Hearing Review Aug 16, 2017
People who are hearing impaired sometimes have a difficult time following a conversation in a multi–speaker environment such as a noisy restaurant or a party. While current hearing aids can suppress background noise, users still have difficulty listening to a single conversation among many without knowing which speaker the user is attending to. A cognitive hearing aid that constantly monitors the brain activity of the subject to determine whether the subject is conversing with a specific speaker in the environment would be a dream come true.
Using deep neural network models, researchers at Columbia Engineering have made a breakthrough in auditory attention decoding (AAD) methods and are coming closer to making cognitively–controlled hearing aids a reality, the university announced on its website. The study, led by Nima Mesgarani, associate professor of electrical engineering, was published in the Journal of Neural Engineering. The work was done in collaboration with Columbia University Medical CenterÂs Department of Neurosurgery, Hofstra–Northwell School of Medicine, and Feinstein Institute for Medical Research. DEMO: http://naplab.ee.columbia.edu/nnaad.html
MesgaraniÂs team developed an end–to–end system that receives a single audio channel containing a mixture of speakers by a listener along with the listenerÂs neural signals, automatically separates the individual speakers in the mixture, determines which speaker is being listened to, and then amplifies the attended speakerÂs voice to assist the listener – all in under 10 seconds.
ÂTranslating these findings to real–world applications poses many challenges, notes James OÂSullivan, a postdoctoral research scientist working with Mesgarani and lead author of the study. According to Columbia Engineering, in a typical implementation of auditory attention decoding, researchers compare the neural responses recorded from a subjectÂs brain with the clean speech uttered by different speakers; the speaker who produces the maximum similarity with the neural data is determined to be the target and is subsequently amplified. However, in the real world, researchers have access only to the mixture, not the individual speakers.
ÂOur study takes a significant step towards automatically separating an attended speaker from the mixture, OÂSullivan said.ÂTo do so, we built deep neural network models that can automatically separate specific speakers from a mixture. We then compare each of these separated speakers with the neural signals to determine which voice the subject is listening to, and then amplify that specific voice for the listener.Â
The team tested the efficacy of their system using invasive electrocorticography recordings from neurological subjects undergoing epilepsy surgery. They identified the regions of the auditory cortex that contribute to AAD and found that the system decoded the attention of the listener and amplified the voice he or she wanted to listen to, using only the mixed audio.
ÂOur system demonstrates a significant improvement in both subjective and objective speech quality measures – almost all of our subjects said they wanted to continue to use it, Mesgarani said. ÂOur novel framework for AAD bridges the gap between the most recent advancements in speech processing technologies and speech prosthesis research and moves us closer to the development of realistic hearing aid devices that can automatically and dynamically track a userÂs direction of attention and amplify an attended speaker.Â
Go to Original
Using deep neural network models, researchers at Columbia Engineering have made a breakthrough in auditory attention decoding (AAD) methods and are coming closer to making cognitively–controlled hearing aids a reality, the university announced on its website. The study, led by Nima Mesgarani, associate professor of electrical engineering, was published in the Journal of Neural Engineering. The work was done in collaboration with Columbia University Medical CenterÂs Department of Neurosurgery, Hofstra–Northwell School of Medicine, and Feinstein Institute for Medical Research. DEMO: http://naplab.ee.columbia.edu/nnaad.html
MesgaraniÂs team developed an end–to–end system that receives a single audio channel containing a mixture of speakers by a listener along with the listenerÂs neural signals, automatically separates the individual speakers in the mixture, determines which speaker is being listened to, and then amplifies the attended speakerÂs voice to assist the listener – all in under 10 seconds.
ÂTranslating these findings to real–world applications poses many challenges, notes James OÂSullivan, a postdoctoral research scientist working with Mesgarani and lead author of the study. According to Columbia Engineering, in a typical implementation of auditory attention decoding, researchers compare the neural responses recorded from a subjectÂs brain with the clean speech uttered by different speakers; the speaker who produces the maximum similarity with the neural data is determined to be the target and is subsequently amplified. However, in the real world, researchers have access only to the mixture, not the individual speakers.
ÂOur study takes a significant step towards automatically separating an attended speaker from the mixture, OÂSullivan said.ÂTo do so, we built deep neural network models that can automatically separate specific speakers from a mixture. We then compare each of these separated speakers with the neural signals to determine which voice the subject is listening to, and then amplify that specific voice for the listener.Â
The team tested the efficacy of their system using invasive electrocorticography recordings from neurological subjects undergoing epilepsy surgery. They identified the regions of the auditory cortex that contribute to AAD and found that the system decoded the attention of the listener and amplified the voice he or she wanted to listen to, using only the mixed audio.
ÂOur system demonstrates a significant improvement in both subjective and objective speech quality measures – almost all of our subjects said they wanted to continue to use it, Mesgarani said. ÂOur novel framework for AAD bridges the gap between the most recent advancements in speech processing technologies and speech prosthesis research and moves us closer to the development of realistic hearing aid devices that can automatically and dynamically track a userÂs direction of attention and amplify an attended speaker.Â
Only Doctors with an M3 India account can read this article. Sign up for free or login with your existing account.
4 reasons why Doctors love M3 India
-
Exclusive Write-ups & Webinars by KOLs
-
Daily Quiz by specialty
-
Paid Market Research Surveys
-
Case discussions, News & Journals' summaries