We propose that using a continuous trajectory model to describe an articulatory-based feature set will address some of the shortcomings inherent in the hidden Markov model (HMM) as a model for speech recognition. The articulatory parameters allow us to explicitly model effects such as co-articulation and assimilation. A linear dynamic model (LDM) is used to capture the characteristics of each segment type. These models are well suited to describing smoothly varying, continuous, yet noisy trajectories, such as we find present in speech data. Experimentation has been based on data for a single speaker from the MOCHA corpus. This consists of parallel acoustic and recorded articulatory parameters for 460 TIMIT sentences. We report the results o...
Abstract Most of the current state-of-the-art speech recognition systems are based on speech signal ...
This paper presents an investigation of ways to integrate articulatory features into Hidden Markov M...
In this paper we present a method to predict the movement of a speaker's mouth from text input using...
The majority of automatic speech recognition (ASR) systems rely on hidden Markov models (HMM), in wh...
We describe the modelling of articulatory movements using (hidden) dynamical system models trained o...
We describe the modelling of articulatory movements using (hidden) dynamical system models trained o...
In this letter, we introduce an hidden Markov model (HMM)-based inversion system to recovery articul...
Institute for Communicating and Collaborative SystemsThe conditional independence assumption imposed...
In this paper we investigate the use of articulatory data for speech recognition. Recordings of the ...
The conditional independence assumption imposed by the hidden Markov models (HMMs) makes it difficul...
A Hidden-Articulator Markov Model (HAMM) is a Hidden Markov Model (HMM) in which each state represen...
The majority of automatic speech recognition (ASR) systems rely on hidden Markov models, in which Ga...
We describe a dynamic Bayesian network for articulatory feature recognition. The model is intended t...
In the past decades, statistics-based hidden Markov models (HMMs) have become the predominant approa...
This paper describes the use of dynamic Bayesian networks for the task of articulatory feature recog...
Abstract Most of the current state-of-the-art speech recognition systems are based on speech signal ...
This paper presents an investigation of ways to integrate articulatory features into Hidden Markov M...
In this paper we present a method to predict the movement of a speaker's mouth from text input using...
The majority of automatic speech recognition (ASR) systems rely on hidden Markov models (HMM), in wh...
We describe the modelling of articulatory movements using (hidden) dynamical system models trained o...
We describe the modelling of articulatory movements using (hidden) dynamical system models trained o...
In this letter, we introduce an hidden Markov model (HMM)-based inversion system to recovery articul...
Institute for Communicating and Collaborative SystemsThe conditional independence assumption imposed...
In this paper we investigate the use of articulatory data for speech recognition. Recordings of the ...
The conditional independence assumption imposed by the hidden Markov models (HMMs) makes it difficul...
A Hidden-Articulator Markov Model (HAMM) is a Hidden Markov Model (HMM) in which each state represen...
The majority of automatic speech recognition (ASR) systems rely on hidden Markov models, in which Ga...
We describe a dynamic Bayesian network for articulatory feature recognition. The model is intended t...
In the past decades, statistics-based hidden Markov models (HMMs) have become the predominant approa...
This paper describes the use of dynamic Bayesian networks for the task of articulatory feature recog...
Abstract Most of the current state-of-the-art speech recognition systems are based on speech signal ...
This paper presents an investigation of ways to integrate articulatory features into Hidden Markov M...
In this paper we present a method to predict the movement of a speaker's mouth from text input using...