In recent years, there have been numerous developments toward solving multimodal tasks, aiming to learn a stronger representation than through a single modality. Certain aspects of the data can be particularly useful in this case--for example, correlations in the space or time domain across modalities--but should be wisely exploited in order to benefit from their full predictive potential. We propose two deep learning architectures with multimodal cross connections that allow for dataflow between several feature extractors (XFlow). Our models derive more interpretable features and achieve better performances than models that do not exchange representations, usefully exploiting correlations between audio and visual data, which have a differe...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...
In recent years, there have been numerous developments towards solving multimodal tasks, aiming to l...
In this paper, we propose a multimodal deep learning architecture for emotion recognition in video r...
Comunicació presentada a: 18th International Society for Music Information Retrieval Conference (ISM...
In this paper, we propose a multimodal deep learning architecture for emotion recognition in video r...
International audienceIn this paper, we propose a multimodal deep learning architecturefor emotion r...
This work aims at investigating cross-modal connections between audio and video sources in the task ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
This work aims at investigating cross-modal connections between audio and video sources in the task ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...
In recent years, there have been numerous developments towards solving multimodal tasks, aiming to l...
In this paper, we propose a multimodal deep learning architecture for emotion recognition in video r...
Comunicació presentada a: 18th International Society for Music Information Retrieval Conference (ISM...
In this paper, we propose a multimodal deep learning architecture for emotion recognition in video r...
International audienceIn this paper, we propose a multimodal deep learning architecturefor emotion r...
This work aims at investigating cross-modal connections between audio and video sources in the task ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
This work aims at investigating cross-modal connections between audio and video sources in the task ...
We propose a novel deep training algorithm for joint representation of audio and visual information ...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...
International audienceHuman affects and automatic emotions detection has been an active research top...