In this study, we present an innovative technique for speaker adaptation in order to improve the accuracy of segmentation with application to unit-selection Text-To-Speech (TTS) systems. Unlike conventional techniques for speaker adaptation, which attempt to improve the accuracy of the segmentation using acoustic models that are more robust in the face of the speaker\u27s characteristics, we aim to use only context dependent characteristics extrapolated with linguistic analysis techniques. In simple terms, we use the intuitive idea that context dependent information is tightly correlated with the related acoustic waveform. We propose a statistical model, which predicts correcting values to reduce the systematic error produced by a state-of-...
The EMIME European project is conducting research in the development of technologies for mobile, per...
This paper presents improved HMM/SVM methods for a two-stage phoneme segmentation framework, which t...
ICASSP2006: IEEE International Conference on Acoustics, Speech, and Signal Processing, May 14-19, ...
Speech segmentation refers to the problem of determining the phoneme boundaries from an acoustic rec...
In the past few years numerous techniques have been proposed to improve the efficiency of basic adap...
This paper describes the refinement of the automatic speech segmentation into phones obtained via Hi...
Orientador: Fabio ViolaroTese (doutorado) - Universidade Estadual de Campinas, Faculdade de Engenhar...
Approaches in Automatic Speech Recognition based on classic acoustic models seem not to...
Determination of the error rate In the speech database collected here mostly in 1995, there are cur...
Phonetic segmentation is the breakup and classication of the sound signal into a string of phones. T...
Correct and temporally accurate phonetic segmentation of speech utterances is important in applicati...
Abstract—In this paper, we propose a robust compensation strategy to deal effectively with extraneou...
For segmenting a speech database, using a family of acoustic models provides multiple estimates of e...
www.talp.upc.es In the present paper we present two novel approaches to phonetic speech segmentation...
This thesis describes work developing an approach to automatic speech recognition which incorporates...
The EMIME European project is conducting research in the development of technologies for mobile, per...
This paper presents improved HMM/SVM methods for a two-stage phoneme segmentation framework, which t...
ICASSP2006: IEEE International Conference on Acoustics, Speech, and Signal Processing, May 14-19, ...
Speech segmentation refers to the problem of determining the phoneme boundaries from an acoustic rec...
In the past few years numerous techniques have been proposed to improve the efficiency of basic adap...
This paper describes the refinement of the automatic speech segmentation into phones obtained via Hi...
Orientador: Fabio ViolaroTese (doutorado) - Universidade Estadual de Campinas, Faculdade de Engenhar...
Approaches in Automatic Speech Recognition based on classic acoustic models seem not to...
Determination of the error rate In the speech database collected here mostly in 1995, there are cur...
Phonetic segmentation is the breakup and classication of the sound signal into a string of phones. T...
Correct and temporally accurate phonetic segmentation of speech utterances is important in applicati...
Abstract—In this paper, we propose a robust compensation strategy to deal effectively with extraneou...
For segmenting a speech database, using a family of acoustic models provides multiple estimates of e...
www.talp.upc.es In the present paper we present two novel approaches to phonetic speech segmentation...
This thesis describes work developing an approach to automatic speech recognition which incorporates...
The EMIME European project is conducting research in the development of technologies for mobile, per...
This paper presents improved HMM/SVM methods for a two-stage phoneme segmentation framework, which t...
ICASSP2006: IEEE International Conference on Acoustics, Speech, and Signal Processing, May 14-19, ...