International audienceWe present a joint spatial and spectral denoising front-end for Track 1 of the 2nd CHiME Speech Separation and Recognition Challenge based on the Flexible Audio Source Separation Toolbox (FASST). We represent the sources by nonnegative matrix factorization (NMF) and full-rank spatial covariances, which are known to be appropriate for the modeling of small source movements. We then learn acoustic models for automatic speech recognition (ASR) on the enhanced training data. We obtain 40% average error rate reduction due to speech separation compared to multicondition training alone
Communication by speech is intrinsic for humans. Since the breakthrough of mobile devices and wirele...
Submitted to ICASSP 2020International audienceWe consider the problem of robust automatic speech rec...
This thesis takes the classical signal processing problem of separating the speech of a target speak...
We present a joint spatial and spectral denoising front-end for Track 1 of the 2nd CHiME Speech Sepa...
International audienceWe describe our submission to the 2011 CHiME Speech Separation and Recognition...
International audienceDistant-microphone automatic speech recognition (ASR) remains a challenging go...
International audienceWe consider the FASST framework for audio source separation, which models the ...
International audienceMulti-microphone signal processing techniques have the potential to greatly im...
International audienceThe CHiME challenge series aims to advance robust automatic speech recognition...
International audienceDistant microphone speech recognition systems that operate with humanlike robu...
International audienceThis article addresses the modeling of reverberant recording environments in t...
International audienceThe CHiME challenge series aims to advance far field speech recognition techno...
International audienceDistant-microphone automatic speech recognition (ASR) remains a challenging go...
The majority of automatic speech recognition (ASR) systems rely on hidden Markov models, in which G...
When speech is captured with a distant microphone, it includes distortions caused by noise, reverber...
Communication by speech is intrinsic for humans. Since the breakthrough of mobile devices and wirele...
Submitted to ICASSP 2020International audienceWe consider the problem of robust automatic speech rec...
This thesis takes the classical signal processing problem of separating the speech of a target speak...
We present a joint spatial and spectral denoising front-end for Track 1 of the 2nd CHiME Speech Sepa...
International audienceWe describe our submission to the 2011 CHiME Speech Separation and Recognition...
International audienceDistant-microphone automatic speech recognition (ASR) remains a challenging go...
International audienceWe consider the FASST framework for audio source separation, which models the ...
International audienceMulti-microphone signal processing techniques have the potential to greatly im...
International audienceThe CHiME challenge series aims to advance robust automatic speech recognition...
International audienceDistant microphone speech recognition systems that operate with humanlike robu...
International audienceThis article addresses the modeling of reverberant recording environments in t...
International audienceThe CHiME challenge series aims to advance far field speech recognition techno...
International audienceDistant-microphone automatic speech recognition (ASR) remains a challenging go...
The majority of automatic speech recognition (ASR) systems rely on hidden Markov models, in which G...
When speech is captured with a distant microphone, it includes distortions caused by noise, reverber...
Communication by speech is intrinsic for humans. Since the breakthrough of mobile devices and wirele...
Submitted to ICASSP 2020International audienceWe consider the problem of robust automatic speech rec...
This thesis takes the classical signal processing problem of separating the speech of a target speak...