Text representation can map text into a vector space for subsequent use in numerical calculations and processing tasks. Word embedding is an important component of text representation. Most existing word embedding models focus on writing and utilize context, weight, dependency, morphology, etc., to optimize the training. However, from the linguistic point of view, spoken language is a more direct expression of semantics; writing has meaning only as a recording of spoken language. Therefore, this paper proposes the concept of a pronunciation-enhanced word embedding model (PWE) that integrates speech information into training to fully apply the roles of both speech and writing to meaning. This paper uses the Chinese language, English language...
We propose a simple yet effective approach to learning bilingual word embeddings (BWEs) from non-par...
We propose cw2vec, a novel method for learning Chinese word embeddings. It is based on our observati...
We introduce bilingual word embeddings: semantic embeddings associated across two languages in the c...
Most word embedding methods take a word as a ba-sic unit and learn embeddings according to words’ ex...
In this paper we propose a novel word representation for Chinese based on a state-of-the-art word em...
Models for statistical spoken language understanding (SLU) systems are conventionally trained using ...
The most recent end-to-end speech synthesis systems use phonemes as acoustic input tokens and ignore...
Speech recognition systems have used the concept of states as a way to decompose words into sub-word...
Research on word representation has always been an important area of interest in the antiquity of Na...
We propose a new model for learning bilingual word representations from non-parallel document-aligne...
Word embeddings have become ubiquitous in NLP, especially when using neural networks. One of the ass...
Transforming an acoustic signal to words is the gold standard in automatic speech recognition. Whil...
In recent years it has become clear that data is the new resource of power and richness. The compani...
We attemped to improve recognition accuracy by reduc-ing the inadequacies of the lexicon and languag...
Recent work has shown success in learning word embeddings with neural network language models (NNLM)...
We propose a simple yet effective approach to learning bilingual word embeddings (BWEs) from non-par...
We propose cw2vec, a novel method for learning Chinese word embeddings. It is based on our observati...
We introduce bilingual word embeddings: semantic embeddings associated across two languages in the c...
Most word embedding methods take a word as a ba-sic unit and learn embeddings according to words’ ex...
In this paper we propose a novel word representation for Chinese based on a state-of-the-art word em...
Models for statistical spoken language understanding (SLU) systems are conventionally trained using ...
The most recent end-to-end speech synthesis systems use phonemes as acoustic input tokens and ignore...
Speech recognition systems have used the concept of states as a way to decompose words into sub-word...
Research on word representation has always been an important area of interest in the antiquity of Na...
We propose a new model for learning bilingual word representations from non-parallel document-aligne...
Word embeddings have become ubiquitous in NLP, especially when using neural networks. One of the ass...
Transforming an acoustic signal to words is the gold standard in automatic speech recognition. Whil...
In recent years it has become clear that data is the new resource of power and richness. The compani...
We attemped to improve recognition accuracy by reduc-ing the inadequacies of the lexicon and languag...
Recent work has shown success in learning word embeddings with neural network language models (NNLM)...
We propose a simple yet effective approach to learning bilingual word embeddings (BWEs) from non-par...
We propose cw2vec, a novel method for learning Chinese word embeddings. It is based on our observati...
We introduce bilingual word embeddings: semantic embeddings associated across two languages in the c...