Pretrained transformer-based language models achieve state-of-the-art performance in many NLP tasks, but it is an open question whether the knowledge acquired by the models during pretraining resembles the linguistic knowledge of humans. We present both humans and pretrained transformers with descriptions of events, and measure their preference for telic interpretations (the event has a natural endpoint) or atelic interpretations (the event does not have a natural endpoint). To measure these preferences and determine what factors influence them, we design an English test and a novel-word test that include a variety of linguistic cues (noun phrase quantity, resultative structure, contextual information, temporal units) that bias toward certa...
The transformers that drive chatbots and other AI systems constitute large language models (LLMs). T...
Five thousand variations of the RoBERTa model, an artificially intelligent "transformer" that can un...
Transformer based language models exhibit intelligent behaviors such as understanding natural langua...
Deep learning has the potential to help solve numerous problems in cognitive science andeducation, b...
Despite being designed for performance rather than cognitive plausibility, transformer language mode...
Language Generation Models produce words based on the previous context. Although existing methods of...
Features of the physical world may be acquired from the statistical properties of language. Here we ...
Pre-trained transformers have rapidly become very popular in the Natural Language Processing (NLP) c...
International audienceAttention mechanisms have played a crucial role in the development of complex ...
This chapter presents an overview of the state of the art in natural language processing, exploring ...
We analyze the Knowledge Neurons framework for the attribution of factual and relational knowledge t...
The goal of my thesis is to investigate the most influential transformer architectures and to apply ...
Thesis (Master's)--University of Washington, 2021Transformer models perform well on NLP tasks, but r...
We analyze if large language models are able to predict patterns of human reading behavior. We compa...
We analyze if large language models are able to predict patterns of human reading behavior. We compa...
The transformers that drive chatbots and other AI systems constitute large language models (LLMs). T...
Five thousand variations of the RoBERTa model, an artificially intelligent "transformer" that can un...
Transformer based language models exhibit intelligent behaviors such as understanding natural langua...
Deep learning has the potential to help solve numerous problems in cognitive science andeducation, b...
Despite being designed for performance rather than cognitive plausibility, transformer language mode...
Language Generation Models produce words based on the previous context. Although existing methods of...
Features of the physical world may be acquired from the statistical properties of language. Here we ...
Pre-trained transformers have rapidly become very popular in the Natural Language Processing (NLP) c...
International audienceAttention mechanisms have played a crucial role in the development of complex ...
This chapter presents an overview of the state of the art in natural language processing, exploring ...
We analyze the Knowledge Neurons framework for the attribution of factual and relational knowledge t...
The goal of my thesis is to investigate the most influential transformer architectures and to apply ...
Thesis (Master's)--University of Washington, 2021Transformer models perform well on NLP tasks, but r...
We analyze if large language models are able to predict patterns of human reading behavior. We compa...
We analyze if large language models are able to predict patterns of human reading behavior. We compa...
The transformers that drive chatbots and other AI systems constitute large language models (LLMs). T...
Five thousand variations of the RoBERTa model, an artificially intelligent "transformer" that can un...
Transformer based language models exhibit intelligent behaviors such as understanding natural langua...