Researcher uses machine learning to teach a computer program how to remove unwanted noise and enhance speech. Future hearing aid users will be able to target their listening more accurately thanks to new Danish technology.
One of the main challenges for people with hearing loss is understanding speech in noisy surroundings. The problem is referred to as the cocktail party effect because situations where many people are talking at the same time often make it very hard to distinguish what is being said by the individual you are talking to.
Even though most modern hearing aids incorporate various forms of speech enhancement technology, engineers are still struggling to develop a system that makes a significant improvement.
Computer listens and learns
The hearing center inside our brains usually performs a string of wildly complicated calculations that enables us to focus on a single voice even if many other people are talking in the background.
Mathew Kavalekalam started out with a digital model that describes how speech is produced in a human body, from the lungs via throat and larynx, mouth and nasal cavities, teeth, lips, etc. He used the model to describe the type of signal that a computer should 'listen' for when trying to identify a talking voice. He then told the computer to start listening and learning.
Noise isn't just noise
Background noise differs depending on the environment, from the street or traffic noise if you are outside to the noise of people talking in a pub or a cafeteria. That is one of the many reasons why it is so tricky to build a model for speech enhancement that filters the speech you want to hear from the babbling you are not interested in.
At Aalborg University Mathew Kavalekalam played back various recordings of voices talking to the computer and gradually added different types of background noise at an increasing level. By applying this machine learning, the computer software developed a way of recognizing the sound patterns and calculating how to enhance the particular sound of talking voices and not the background noise.
Snappy signal processing
However, there is still some work to be done before Mathew Kavalekalam's software finds its way into new hearing aids. The technology needs to be tweaked and tuned before it is practically applicable. The algorithm needs to be optimized to take up less processing power. Even though technology keeps getting faster and more powerful, there are hardware limitations in small, modern hearing aids.
When it comes to speech enhancement, signal processing needs to be snappy. If the sound is delayed in the hearing aid, it gets out of sync with the mouth movements, and that will end up making you even more confused.
1. One in six Europeans experiences various degrees of hearing impairment. Almost everyone loses part of their hearing as they age.
2. Hearing loss often manifests itself in problems when trying to participate in conversations with more than one person talking. This can lead to isolation as people with hearing loss often choose to withdraw from social gatherings where they have to spend a lot of energy trying to keep up with what is being said.