Language Generation Models produce words based on the previous context. Although existing methods offer input attributions as explanations for a model's prediction, it is still unclear how prior words affect the model's decision throughout the layers. In this work, we leverage recent advances in explainability of the Transformer and present a procedure to analyze models for language generation. Using contrastive examples, we compare the alignment of our explanations with evidence of the linguistic phenomena, and show that our method consistently aligns better than gradient-based and perturbation-based baselines. Then, we investigate the role of MLPs inside the Transformer and show that they learn features that help the model predict words t...
We analyze the Knowledge Neurons framework for the attribution of factual and relational knowledge t...
This chapter presents an overview of the state of the art in natural language processing, exploring ...
Deep learning has the potential to help solve numerous problems in cognitive science andeducation, b...
Language Generation Models produce words based on the previous context. Although existing methods of...
In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and...
In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and...
Pretrained transformer-based language models achieve state-of-the-art performance in many NLP tasks,...
Transformer based language models exhibit intelligent behaviors such as understanding natural langua...
We seek to understand how the representations of individual tokens and the structure of the learned ...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Thesis (Master's)--University of Washington, 2021Transformer models perform well on NLP tasks, but r...
Establishing whether language models can use contextual information in a human-plausible way is impo...
International audienceWe probe pre-trained transformer language models for bridging inference. We fi...
Transformers have been established as one of the most effective neural approach in performing variou...
Transformer is a neural machine translation model which revolutionizes machine translation. Compared...
We analyze the Knowledge Neurons framework for the attribution of factual and relational knowledge t...
This chapter presents an overview of the state of the art in natural language processing, exploring ...
Deep learning has the potential to help solve numerous problems in cognitive science andeducation, b...
Language Generation Models produce words based on the previous context. Although existing methods of...
In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and...
In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and...
Pretrained transformer-based language models achieve state-of-the-art performance in many NLP tasks,...
Transformer based language models exhibit intelligent behaviors such as understanding natural langua...
We seek to understand how the representations of individual tokens and the structure of the learned ...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Thesis (Master's)--University of Washington, 2021Transformer models perform well on NLP tasks, but r...
Establishing whether language models can use contextual information in a human-plausible way is impo...
International audienceWe probe pre-trained transformer language models for bridging inference. We fi...
Transformers have been established as one of the most effective neural approach in performing variou...
Transformer is a neural machine translation model which revolutionizes machine translation. Compared...
We analyze the Knowledge Neurons framework for the attribution of factual and relational knowledge t...
This chapter presents an overview of the state of the art in natural language processing, exploring ...
Deep learning has the potential to help solve numerous problems in cognitive science andeducation, b...