This work tests several classification techniques and acoustic features and further combines them using late fusion to classify paralinguistic information for the ComParE 2018 challenge. We use Multiple Linear Regression (MLR) with Ordinary Least Squares (OLS) analysis to select the most informative features for Self-Assessed Affect (SSA) sub-Challenge. We also propose to use raw-waveform convolutional neural networks (CNN) in the context of three paralinguistic sub-challenges. By using combined evaluation split for estimating codebook, we obtain better representation for Bag-of-Audio-Words approach. We preprocess the speech to vocalized segments to improve classification performance. For fusion of our leading classification techniques, we ...
In this position paper we present the FP7 ERC starting grant project iHEARu (Intelligent systems ’ H...
In this contribution, we investigate the effectiveness of deep fusion of text and audio features for...
We investigate classification of non-linguistic vocalisations with a novel audiovisual approach and ...
Senior Project submitted to The Division of Science, Mathematics and Computing of Bard College. Abst...
17th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2016) -- S...
This document describes the three methods for the detection and classification of paralinguistic exp...
Abstract The INTERSPEECH 2018 Computational Paralinguistics Challenge addresses four different prob...
This paper investigates different fusion strategies as well as provides insights on their effectiven...
International audienceSpeech classifiers of paralinguistic traits traditionally learn from diverse h...
Analysing the voice behind the words represents a central aspect of human communication and is thus...
Abstract This paper describes Brno University of Technology (BUT) system for the Interspeech 2010 Pa...
The automatic analysis of speech to detect affective states may improve the way users interact with ...
18th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2017) -- A...
In this article we propose two algorithms for discourse prosodic feature interpretation. The first a...
Affective computing is becoming increasingly significant in the interaction between humans and machi...
In this position paper we present the FP7 ERC starting grant project iHEARu (Intelligent systems ’ H...
In this contribution, we investigate the effectiveness of deep fusion of text and audio features for...
We investigate classification of non-linguistic vocalisations with a novel audiovisual approach and ...
Senior Project submitted to The Division of Science, Mathematics and Computing of Bard College. Abst...
17th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2016) -- S...
This document describes the three methods for the detection and classification of paralinguistic exp...
Abstract The INTERSPEECH 2018 Computational Paralinguistics Challenge addresses four different prob...
This paper investigates different fusion strategies as well as provides insights on their effectiven...
International audienceSpeech classifiers of paralinguistic traits traditionally learn from diverse h...
Analysing the voice behind the words represents a central aspect of human communication and is thus...
Abstract This paper describes Brno University of Technology (BUT) system for the Interspeech 2010 Pa...
The automatic analysis of speech to detect affective states may improve the way users interact with ...
18th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2017) -- A...
In this article we propose two algorithms for discourse prosodic feature interpretation. The first a...
Affective computing is becoming increasingly significant in the interaction between humans and machi...
In this position paper we present the FP7 ERC starting grant project iHEARu (Intelligent systems ’ H...
In this contribution, we investigate the effectiveness of deep fusion of text and audio features for...
We investigate classification of non-linguistic vocalisations with a novel audiovisual approach and ...