We present a comprehensive study on the effect of reverberation and background noise on the recognition of nonprototypical emotions from speech. We carry out our evaluation on a single, well-defined task based on the FAU Aibo Emotion Corpus consisting of spontaneous children's speech, which was used in the INTERSPEECH 2009 Emotion Challenge, the first of its kind. Based on the challenge task, and relying on well-proven methodologies from the speech recognition domain, we derive test scenarios with realistic noise and reverberation conditions, including matched as well as mismatched condition training. As feature extraction based on supervised Nonnegative Matrix Factorization (NMF) has been proposed in automatic speech recognition for enhanc...
In this paper, we test the use of Nonnegative Matrix Fac-torization (NMF) for feature extraction in ...
Speech is a direct and rich way of transmitting information and emotions from one point to another. ...
This paper presents robust recognition of a subset of emotions by animated agents from salient spoke...
Copyright © 2011 Felix Weninger et al. This is an open access article distributed under the Creative...
In recent times, much research is progressing forward in the field of speech emotion recognition (SE...
MasterRecognizing human emotion from speech signals suffers from uncertainties in both representatio...
This paper investigates the performance of “Deep Learning” for speech emotion classification when th...
This paper describes a novel two-stage dereverberation feature enhancement method for noise-robust a...
Automatic recognition of speech emotional states in noisy conditions has become an important researc...
We present a speech signal driven emotion recognition system. Our system is trained and tested with ...
This work presents an approach for text-independent and speaker-independent emotion recognition from...
The problem of reverberation in speech recognition is addressed in this study by extending a noise-r...
Speech is a direct and rich way of transmitting information and emotions from one point to another. ...
Speech feature fusion is the most commonly used phenomenon for improving the accuracy in Speech Emot...
In one study on vocal emotion recognition using noise-vocoded speech (NVS), the high similarities be...
In this paper, we test the use of Nonnegative Matrix Fac-torization (NMF) for feature extraction in ...
Speech is a direct and rich way of transmitting information and emotions from one point to another. ...
This paper presents robust recognition of a subset of emotions by animated agents from salient spoke...
Copyright © 2011 Felix Weninger et al. This is an open access article distributed under the Creative...
In recent times, much research is progressing forward in the field of speech emotion recognition (SE...
MasterRecognizing human emotion from speech signals suffers from uncertainties in both representatio...
This paper investigates the performance of “Deep Learning” for speech emotion classification when th...
This paper describes a novel two-stage dereverberation feature enhancement method for noise-robust a...
Automatic recognition of speech emotional states in noisy conditions has become an important researc...
We present a speech signal driven emotion recognition system. Our system is trained and tested with ...
This work presents an approach for text-independent and speaker-independent emotion recognition from...
The problem of reverberation in speech recognition is addressed in this study by extending a noise-r...
Speech is a direct and rich way of transmitting information and emotions from one point to another. ...
Speech feature fusion is the most commonly used phenomenon for improving the accuracy in Speech Emot...
In one study on vocal emotion recognition using noise-vocoded speech (NVS), the high similarities be...
In this paper, we test the use of Nonnegative Matrix Fac-torization (NMF) for feature extraction in ...
Speech is a direct and rich way of transmitting information and emotions from one point to another. ...
This paper presents robust recognition of a subset of emotions by animated agents from salient spoke...