Retrieval-augmented language models (LMs) use non-parametric memory to substantially outperform their non-retrieval counterparts on perplexity-based evaluations, but it is an open question whether they achieve similar gains in few- and zero-shot end-task accuracy. We extensively study one such model, the k-nearest neighbor LM (kNN-LM), showing that the gains marginally transfer. The main challenge is to achieve coverage of the verbalizer tokens that define the different end-task class labels. To address this challenge, we also introduce kNN-Prompt, a simple and effective kNN-LM with automatically expanded fuzzy verbalizers (e.g. to expand terrible to also include silly and other task-specific synonyms for sentiment classification). Across n...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
Natural Language Inference (NLI) plays an important role in many natural language processing tasks s...
Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained models with relation triples...
K-Nearest Neighbor Neural Machine Translation (kNN-MT) successfully incorporates external corpus by ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
Most combinations of NLP tasks and language varieties lack in-domain examples for supervised trainin...
k-Nearest-Neighbor Machine Translation (kNN-MT) has been recently proposed as a non-parametric solut...
Semi-parametric models, which augment generation with retrieval, have led to impressive results in l...
NLP has yielded results that were unimaginable only a few years ago on a wide range of real-world ta...
Recently, $k$NN-MT has shown the promising capability of directly incorporating the pre-trained neur...
In this work, we aim to capitalize on the unique few-shot capabilities of large-scale language model...
When pre-trained on large unsupervised textual corpora, language models are able to store and retri...
Neural language models have drastically changed the landscape of natural language processing (NLP). ...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
Natural Language Inference (NLI) plays an important role in many natural language processing tasks s...
Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained models with relation triples...
K-Nearest Neighbor Neural Machine Translation (kNN-MT) successfully incorporates external corpus by ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
Most combinations of NLP tasks and language varieties lack in-domain examples for supervised trainin...
k-Nearest-Neighbor Machine Translation (kNN-MT) has been recently proposed as a non-parametric solut...
Semi-parametric models, which augment generation with retrieval, have led to impressive results in l...
NLP has yielded results that were unimaginable only a few years ago on a wide range of real-world ta...
Recently, $k$NN-MT has shown the promising capability of directly incorporating the pre-trained neur...
In this work, we aim to capitalize on the unique few-shot capabilities of large-scale language model...
When pre-trained on large unsupervised textual corpora, language models are able to store and retri...
Neural language models have drastically changed the landscape of natural language processing (NLP). ...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
Natural Language Inference (NLI) plays an important role in many natural language processing tasks s...
Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained models with relation triples...