Recent efforts in cross-lingual word embedding (CLWE) learning have predominantly focused on fully unsupervised approaches that project monolingual embeddings into a shared cross-lingual space without any cross-lingual signal. The lack of any supervision makes such approaches conceptually attractive. Yet, their only core difference from (weakly) supervised projection-based CLWE methods is in the way they obtain a seed dictionary used to initialize an iterative self-learning procedure. The fully unsupervised methods have arguably become more robust, and their primary use case is CLWE induction for pairs of resource-poor and distant languages. In this paper, we question the ability of even the most robust unsupervised CLWE approaches to induc...
Word embeddings - dense vector representations of a word’s distributional semantics - are an indespe...
Building bilingual lexica from non-parallel data is a long-standing natural language processing rese...
Distributed representations of words which map each word to a continuous vector have proven useful i...
Bilingual Word Embeddings (BWEs) are one of the cornerstones of cross-lingual transfer of NLP models...
Recent research has discovered that a shared bilingual word embedding space can be induced by projec...
Effective projection-based cross-lingual word embedding (CLWE) induction critically relies on the it...
Cross-lingual word embeddings are an increasingly important reseource in cross-lingual methods for N...
Cross-lingual word embeddings are an increasingly important reseource in cross-lingual methods for N...
Unsupervised machine translation---i.e., not assuming any cross-lingual supervision signal, whether ...
Cross-Lingual Word Embeddings (CLWEs) encode words from two or more languages in a shared high-dimen...
We propose a simple yet effective approach to learning bilingual word embeddings (BWEs) from non-par...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Cross-lingual word embeddings aim to bridge the gap between high-resource and low-resource languages...
Word embeddings - dense vector representations of a word’s distributional semantics - are an indespe...
Building bilingual lexica from non-parallel data is a long-standing natural language processing rese...
Distributed representations of words which map each word to a continuous vector have proven useful i...
Bilingual Word Embeddings (BWEs) are one of the cornerstones of cross-lingual transfer of NLP models...
Recent research has discovered that a shared bilingual word embedding space can be induced by projec...
Effective projection-based cross-lingual word embedding (CLWE) induction critically relies on the it...
Cross-lingual word embeddings are an increasingly important reseource in cross-lingual methods for N...
Cross-lingual word embeddings are an increasingly important reseource in cross-lingual methods for N...
Unsupervised machine translation---i.e., not assuming any cross-lingual supervision signal, whether ...
Cross-Lingual Word Embeddings (CLWEs) encode words from two or more languages in a shared high-dimen...
We propose a simple yet effective approach to learning bilingual word embeddings (BWEs) from non-par...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Cross-lingual word embeddings aim to bridge the gap between high-resource and low-resource languages...
Word embeddings - dense vector representations of a word’s distributional semantics - are an indespe...
Building bilingual lexica from non-parallel data is a long-standing natural language processing rese...
Distributed representations of words which map each word to a continuous vector have proven useful i...