Distributed word representations are very useful for capturing semantic information and have been successfully applied in a variety of NLP tasks, especially on English. In this work, we innovatively develop two component-enhanced Chinese character embedding models and their bigram extensions. Distinguished from English word embeddings, our models explore the compositions of Chinese characters, which often serve as semantic indictors inherently. The evaluations on both word similarity and text classification demonstrate the effectiveness of our models. ? 2015 Association for Computational Linguistics.EI829-83
International audienceChinese characters have a complex and hierarchical graphical structure carryin...
In recent years, many scholars have chosen to use word lexicons to incorporate word information into...
This thesis looks into the problem of learning Chinese characters for foreign language learners and ...
In this paper we propose a novel word representation for Chinese based on a state-of-the-art word em...
Most word embedding methods take a word as a ba-sic unit and learn embeddings according to words’ ex...
In the Chinese language, words consist of characters each of which is composed of one or more compon...
Distributional Similarity has attracted considerable attention in the field of natural language proc...
We propose cw2vec, a novel method for learning Chinese word embeddings. It is based on our observati...
Distributional Similarity has attracted considerable attention in the field of natural language proc...
Chinese characters have semantic-rich compositional information in radical form. While almost all pr...
Recent work has shown success in learning word embeddings with neural network language models (NNLM)...
Named Entity Recognition (NER) is an essential part of many natural language processing (NLP) tasks....
This paper presents a truly full character-level neural dependency parser together with a newly rele...
Chinese character decomposition has been used as a feature to enhance Machine Translation (MT) model...
This thesis deals with Chinese characters (Hanzi): their key characteristics and how they could be u...
International audienceChinese characters have a complex and hierarchical graphical structure carryin...
In recent years, many scholars have chosen to use word lexicons to incorporate word information into...
This thesis looks into the problem of learning Chinese characters for foreign language learners and ...
In this paper we propose a novel word representation for Chinese based on a state-of-the-art word em...
Most word embedding methods take a word as a ba-sic unit and learn embeddings according to words’ ex...
In the Chinese language, words consist of characters each of which is composed of one or more compon...
Distributional Similarity has attracted considerable attention in the field of natural language proc...
We propose cw2vec, a novel method for learning Chinese word embeddings. It is based on our observati...
Distributional Similarity has attracted considerable attention in the field of natural language proc...
Chinese characters have semantic-rich compositional information in radical form. While almost all pr...
Recent work has shown success in learning word embeddings with neural network language models (NNLM)...
Named Entity Recognition (NER) is an essential part of many natural language processing (NLP) tasks....
This paper presents a truly full character-level neural dependency parser together with a newly rele...
Chinese character decomposition has been used as a feature to enhance Machine Translation (MT) model...
This thesis deals with Chinese characters (Hanzi): their key characteristics and how they could be u...
International audienceChinese characters have a complex and hierarchical graphical structure carryin...
In recent years, many scholars have chosen to use word lexicons to incorporate word information into...
This thesis looks into the problem of learning Chinese characters for foreign language learners and ...