Self-supervised large language models (LMs) have become a highly-influential and foundational tool for many NLP models. For this reason, their expressivity is an important topic of study. In near-universal practice, given the language context, the model predicts a word from the vocabulary using a single embedded vector representation of both context and dictionary entries. Note that the context sometimes implies that the distribution over predicted words should be multi-modal in embedded space. However, the context’s single-vector representation provably fails to capture such a distribution. To address this limitation, we propose to represent context with multiple vector embeddings, which we term facets. This is distinct from previous work ...
Thesis (Ph.D.)--University of Washington, 2022A robust language processing machine should be able to...
Word embedding models typically learn two types of vectors: target word vectors and context word vec...
Substantial progress has been made in the field of natural language processing (NLP) due to the adve...
This thesis presents new methods for unsupervised learning of distributed representations of words a...
Context-predicting models (more commonly known as embeddings or neural language models) are the new ...
Distributed word embeddings have yielded state-of-the-art performance in many NLP tasks, mainly due ...
Context-predicting models (more com-monly known as embeddings or neural language models) are the new...
Real-valued word embeddings have transformed natural language processing (NLP) applications, recogni...
Word embedding is a feature learning technique which aims at mapping words from a vocabulary into ve...
Semantic vectors associated with the paper "Don't count, predict! A systematic comparison of context...
The semantic relatedness of words has two key dimensions: it can be based on taxonomic information o...
Title from PDF of title page viewed June 14, 2021Thesis advisor: Yugyung LeeVitaIncludes bibliograph...
Distributed word representations have been widely used and proven to be useful in quite a few natura...
We propose a framework to modularize the training of neural language models that use diverse forms o...
When the field of natural language processing (NLP) entered the era of deep neural networks, the tas...
Thesis (Ph.D.)--University of Washington, 2022A robust language processing machine should be able to...
Word embedding models typically learn two types of vectors: target word vectors and context word vec...
Substantial progress has been made in the field of natural language processing (NLP) due to the adve...
This thesis presents new methods for unsupervised learning of distributed representations of words a...
Context-predicting models (more commonly known as embeddings or neural language models) are the new ...
Distributed word embeddings have yielded state-of-the-art performance in many NLP tasks, mainly due ...
Context-predicting models (more com-monly known as embeddings or neural language models) are the new...
Real-valued word embeddings have transformed natural language processing (NLP) applications, recogni...
Word embedding is a feature learning technique which aims at mapping words from a vocabulary into ve...
Semantic vectors associated with the paper "Don't count, predict! A systematic comparison of context...
The semantic relatedness of words has two key dimensions: it can be based on taxonomic information o...
Title from PDF of title page viewed June 14, 2021Thesis advisor: Yugyung LeeVitaIncludes bibliograph...
Distributed word representations have been widely used and proven to be useful in quite a few natura...
We propose a framework to modularize the training of neural language models that use diverse forms o...
When the field of natural language processing (NLP) entered the era of deep neural networks, the tas...
Thesis (Ph.D.)--University of Washington, 2022A robust language processing machine should be able to...
Word embedding models typically learn two types of vectors: target word vectors and context word vec...
Substantial progress has been made in the field of natural language processing (NLP) due to the adve...