This paper examines the degree of correlation between lip and jaw configuration and speech acoustics. The lip and jaw positions are characterised by a system of measurements taken from video images of the speakers face and profile, and the acoustics are represented using line spectral pair parameters and a measure of RMS energy. A correlation is found between the measured acoustic parameters and a linear estimate of the acoustics recovered from the visual data. This correlation exists despite the simplicity of the visual representation and is in rough agreement with correlations measured in earlier work by Yehia et al. using different techniques. However, analysis of the estimation errors suggests that the visual information, as parameteris...
Several studies in the past have shown that the features based on the kinematics of speech articulat...
It is well established that speech perception is improved when we are able to see the speaker talkin...
We used Auditory/visual masks to investigate how the availability of speech signals governs speech p...
This paper examines the degree of correlation between lip and jaw conguration and speech acoustics. ...
This paper investigates the statistical relationship between acoustic and visual speech features for...
The aim of this work is to investigate a selection of audio and visual speech features with the aim ...
The McGurk effect shows in an obvious manner that visual information from a speaker’s articulatory m...
The long term goal of our work is to predict visual confusion matrices from physical measurements. I...
The fine temporal structure of relations of acoustic and visual features has been investigated to im...
<p>A - Top, Inter-lip distance and the auditory envelope for a single 20 second segment from a singl...
The aim of this work is to examine the correlation between audio and visual speech features. The mot...
International audienceLip reading is the ability to partially understand speech by looking at the sp...
Two experiments aimed to determine whether features of both the visual and acoustical inputs are alw...
The role of audio–visual speech synchrony for speaker diarisation is investigated on the multiparty ...
When someone speaks, linguistically relevant movements are produced. Most of the previous work that ...
Several studies in the past have shown that the features based on the kinematics of speech articulat...
It is well established that speech perception is improved when we are able to see the speaker talkin...
We used Auditory/visual masks to investigate how the availability of speech signals governs speech p...
This paper examines the degree of correlation between lip and jaw conguration and speech acoustics. ...
This paper investigates the statistical relationship between acoustic and visual speech features for...
The aim of this work is to investigate a selection of audio and visual speech features with the aim ...
The McGurk effect shows in an obvious manner that visual information from a speaker’s articulatory m...
The long term goal of our work is to predict visual confusion matrices from physical measurements. I...
The fine temporal structure of relations of acoustic and visual features has been investigated to im...
<p>A - Top, Inter-lip distance and the auditory envelope for a single 20 second segment from a singl...
The aim of this work is to examine the correlation between audio and visual speech features. The mot...
International audienceLip reading is the ability to partially understand speech by looking at the sp...
Two experiments aimed to determine whether features of both the visual and acoustical inputs are alw...
The role of audio–visual speech synchrony for speaker diarisation is investigated on the multiparty ...
When someone speaks, linguistically relevant movements are produced. Most of the previous work that ...
Several studies in the past have shown that the features based on the kinematics of speech articulat...
It is well established that speech perception is improved when we are able to see the speaker talkin...
We used Auditory/visual masks to investigate how the availability of speech signals governs speech p...