This technology separates multi-source audio signals and filters the signals based on the listener’s brain activity.
Speech perception in crowded environments is challenging particularly for hearing-impaired listeners. Assistive hearing devices have seen substantial progress in suppressing background noises that are acoustically different from speech, but they cannot enhance a target speaker without knowing with whom the listener is conversing.
This technology is a system for auditory attention decoding in which the brainwaves of listeners are compared with sound sources to determine the attended source, which can then be amplified to facilitate hearing. In realistic situations, however, only mixed audio is available. The inventors created a speech separation algorithm to automatically separate speakers in mixed audio, with no need for the speakers to have prior training. The results have shown that auditory attention decoding with automatically separated speakers is as accurate and fast as using clean speech sounds. The proposed technology significantly improves the subjective and objective quality of the attended speaker. The technology addresses a major obstacle in actualization of auditory attention decoding that can assist hearing-impaired listeners and reduce listening effort for normal-hearing subjects.
This technology has been demonstrated to cleanly separate and amplify a target speaker relative to the interfering sources. Learn more here.
Han, C., O’Sullivan, J., Luo, Y., Herrero, J., Mehta, A., Mesgarani, N., Speaker independent auditory attention decoding without access to clean sources, Science Advances, 2019. / Demo / Press Release / In the news
Luo, Y., Mesgarani, N., Conv-TasNet: surpassing ideal time-frequency magnitude masking for speech separation, IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2019 / Audio Samples
IR CU17276, CU18145, CU16343
Licensing Contact: Beth Kauderer