• No results found

Multi-microphone speaker separation for neuro-steered hearing aids: neural networks versus linear methods

N/A
N/A
Protected

Academic year: 2021

Share "Multi-microphone speaker separation for neuro-steered hearing aids: neural networks versus linear methods"

Copied!
1
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

24 Sep 2019 Page 1/1

Multi-microphone speaker separation for neuro-steered

hearing aids: neural networks versus linear methods

Neetha Das, Jeroen Zegers, Hugo Van Hamme, Tom Francart, Alexander Bertrand

Neuro-steered noise suppression in hearing devices consists of two main parts: separating and denoising the speech sources in the microphone recordings, and detection of the user’s attention. We compared three multi-microphone approaches for the speaker separation and noise-suppression part in such a pipeline: a (linear) multi-channel Wiener filter (MWF), a (non-linear) deep neural network (DNN), or a combination of both where the output of the DNN is used to inform the MWF. The separated speech streams are then used for an EEG-based auditory attention decoding (AAD) module, the result of which decides which audio stream to present to the user. We evaluate the pipeline for 4 different speaker positions and 3 noise conditions and present the improvement in signal-to-noise ratios (SNRs) and AAD accuracies for the three approaches under the different acoustic conditions. While both methods perform well in the easier acoustic conditions, we found that the DNN-only and the DNN-MWF pipeline result in a more robust performance in the more challenging conditions. Finally, we show that the use of more than 1 microphone for source separation has a substantial benefit on the AAD performance over a single-microphone speaker separation.

Acknowledgements

The work is funded by KU Leuven Special Research Fund C14/16/057 and OT/14/119, FWO project nrs. 1.5.123.16N and G0A4918N, FWO SB PhD grant awarded to Jeroen Zegers (1S66217N), and the ERC (637424 and 802895) under the European Union’s Horizon 2020 research and innovation programme.

Referenties

GERELATEERDE DOCUMENTEN

It was previously proven that a binaural noise reduction procedure based on the Speech Distortion Weighted Multi-channel Wiener Filter (SDW-MWF) indeed preserves the speech

 Combined technique: more robust when VAD fails, better performance than fixed beamformers in other scenarios. • Acoustic transfer function estimation

Klasen, 1 Simon Doclo, 1,2 Tim Van den Bogaert, 1 Marc Moonen, 2 Jan Wouters. 1 KU Leuven ESAT, Kasteelpark Arenberg 10, Leuven 2 KU

• Spatial directivity patterns for non-robust and robust beamformer in case of no position errors and small position errors: [0.002 –0.002 0.002] m. Design, implementation,

• Spatial pre-processor and adaptive stage rely on assumptions (e.g. no microphone mismatch, no reverberation,…). • In practice, these assumptions are often

o Multi-channel Wiener filter (but also e.g. Transfer Function GSC) speech cues are preserved noise cues may be distorted. • Preservation of binaural

o Independent processing of left and right hearing aid o Localisation cues are distorted. RMS error per loudspeaker when accumulating all responses of the different test conditions

• Combination of a-priori knowledge and on-line estimation of both speech and noise terms anticipated to enhance robustness.