(page under construction)

Introduction
Today, over 1.5 billion people – nearly 20% of the global population – experience hearing loss, with at least 430 million facing moderate to severe cases requiring intervention. By 2050, it is expected that 700 million people – or 1 in every 10 people – will have disabling hearing loss (hearing loss greater than 35 decibels in the better hearing ear). Hearing aids (HAs) play a vital role in improving quality of life for these individuals by amplifying sound pressure levels and improving the signal-to-noise ratio (SNR) in most settings.
While modern HAs fulfill basic hearing needs, they often fall short when it comes to spatial hearing – the ability to locate where sounds are coming from in a three-dimensional space. This is a critical aspect of natural hearing that helps with communication and safety in everyday environments. To meet this challenge, we’re exploring how machine learning can push hearing technology forward – making hearing aids smarter, faster to develop and better suited to individual needs.
How it works
Hearing aids address hearing loss by amplifying sound and enhancing SNR, but the human auditory system uses binaural cues – differences in signals received by both ears – that allow the brain to localise sound sources and create a sense of acoustic space, thus perceiving sound spatially. Binaural technologies in modern HAs attempt to replicate these cues to offer an immersive listening experience, reconstructing spatial hearing through sources accurately positioned in space. Better spatialisation of speakers and acoustic environments can improve speech intelligibility for hearing-impaired individuals.
To evaluate how well new algorithms preserve partly subjective spatial cues, traditional methods rely on expensive and time-consuming listening tests with human subjects, encompassing a lengthy process that includes obtaining ethical approval, recruiting participants, and conducting individual listening experiments. Machine learning offers a promising alternative. Instead of relying entirely on human testing, spatial perception models which provide objective metrics that can simulate human auditory localisation can be used to train and validate machine learning models.
For instance, the 3D localisation model we have built, shown in Figure 1, bypasses the manual extraction of auditory cues. Instead, it uses a two-step process: (1) calculating the firing rate dataset from auditory nerve fibers (ANF) and medial superior olive (MSO), and (2) predicting sound angles via a deep neural network (DNN). The model estimates the azimuth angle using MSO and ANF firing rates, and the elevation angle via the positive spectral gradient (PSG) of ANF activity. The localisation model incorporates the peripheral processing method, the MSO model and the DNN to emulate sound source directions.

Figure 1. Overview of the proposed auditory localisation model.
Binaural signals are first processed by a peripheral auditory model to generate ANF spikes, which are then converted into MSO spikes using a spiking auditory neuron model. A DNN predicts lateral and polar angles directly from the firing rates of MSO and ANF, respectively.
What are the applications
Using machine learning in acoustics presents many important benefits, especially when applied to hearing aids. For example, enhancing spatial hearing can boost the clarity and intelligibility of speech in complex environments, thus improving HA performance.
Utilising objective spatial perception models to reduce reliance on lengthy human listening tests greatly expedites the innovation process and decreases developmental costs.
Another application is that of personalised hearing technologies, as models that account for hearing loss can support tailored solutions for both normal-hearing and hearing-impaired users. Furthermore, incorporating these models into consumer audio technology could lead to more immersive virtual or augmented reality experiences.
What we are working on
This project – SPHERE (Spatial Hearing Models for Hearing Instruments) – is a continuation of a longstanding research collaboration between EPFL and Sonova, supported by Innosuisse.
Our 3D localisation model is suitable for both normal hearing and hearing-impaired listeners. Unlike many existing models, ours is not limited to the median plane, accounts for hearing loss and supports dynamic auditory cues. Regarding distance perception, the majority of models focus on achieving precise distance estimations from binaural signals rather than mimicking actual human behavior. We aim to develop a model that reflects real human behavior and incorporates the effects of hearing loss. As for externalization models, current iterations fail to consider the influence of hearing loss and dynamic situations. We are working on an advanced externalization model that integrates static auditory cues while also accommodating hearing loss and dynamic cues. Once these models are constructed, they will first be validated using existing open-source data. The corresponding listening tests will be subsequently designed and conducted based on the preliminary results.
List of projects associated with this topic
| Funding body | Project | Period |
|
CTI |
Binaural Hearing-Aids with Localization and Spatialization BHA(L&S) | 2012-2014 |
|
Innosuisse |
Spatial Hearing Models for Hearing Instruments (SPHERE) | 2023-2025 |

