International audienceComputing pairwise word semantic similarity is widely used and serves as a building block in many tasks in NLP. In this paper, we explore the embedding of the shortest-path metrics from a knowledge base (Wordnet) into the Hamming hyper-cube, in order to enhance the computation performance. We show that, although an isometric embedding is untractable, it is possible to achieve good non-isometric embeddings. We report a speedup of three orders of magnitude for the task of computing Leacock and Chodorow (LCH) similarity while keeping strong correlations (r = .819, ρ = .826)
International audienceIn this paper we discuss the well-known claim that language analogies yield al...
Modelling semantic similarity plays a fundamental role in lexical semantic applications. A natural w...
Abstract. We survey the emerging area of compression-based, parameter-free, similarity distance meas...
International audienceComputing pairwise word semantic similarity is widely used and serves as a bui...
Recent trends suggest that neural-network-inspired word embedding models outperform traditional coun...
Word embeddings seek to recover a Euclidean metric space by mapping words into vectors, starting fro...
Continuous word representations have been remarkably useful across NLP tasks but remain poorly unde...
International audienceWord Embeddings (WE) have recently imposed themselves as a standard for repres...
Word embeddings trained on natural corpora (e.g., newspaper collections, Wikipedia or the Web) excel...
Word embeddings are increasingly attracting the attention of researchers dealing with semantic simil...
Semantic similarity is fundamental operation in the field of computational lexical semantics, artifi...
One of the trends in Natural Language Processing (NLP) is the use of word embedding. Its aim is to b...
We consider the following problem: given neural language models (embeddings) each of which is traine...
AbstractDespite the large diffusion and use of embedding generated through Word2Vec, there are still...
We analyze an approach to a similarity preserving coding of symbol sequences based on neural distrib...
International audienceIn this paper we discuss the well-known claim that language analogies yield al...
Modelling semantic similarity plays a fundamental role in lexical semantic applications. A natural w...
Abstract. We survey the emerging area of compression-based, parameter-free, similarity distance meas...
International audienceComputing pairwise word semantic similarity is widely used and serves as a bui...
Recent trends suggest that neural-network-inspired word embedding models outperform traditional coun...
Word embeddings seek to recover a Euclidean metric space by mapping words into vectors, starting fro...
Continuous word representations have been remarkably useful across NLP tasks but remain poorly unde...
International audienceWord Embeddings (WE) have recently imposed themselves as a standard for repres...
Word embeddings trained on natural corpora (e.g., newspaper collections, Wikipedia or the Web) excel...
Word embeddings are increasingly attracting the attention of researchers dealing with semantic simil...
Semantic similarity is fundamental operation in the field of computational lexical semantics, artifi...
One of the trends in Natural Language Processing (NLP) is the use of word embedding. Its aim is to b...
We consider the following problem: given neural language models (embeddings) each of which is traine...
AbstractDespite the large diffusion and use of embedding generated through Word2Vec, there are still...
We analyze an approach to a similarity preserving coding of symbol sequences based on neural distrib...
International audienceIn this paper we discuss the well-known claim that language analogies yield al...
Modelling semantic similarity plays a fundamental role in lexical semantic applications. A natural w...
Abstract. We survey the emerging area of compression-based, parameter-free, similarity distance meas...