Large language models can store extensive world knowledge, often extractable through question-answering (e.g., "What is Abraham Lincoln's birthday?"). However, it's unclear whether the model answers questions based on exposure to exact/similar questions during training, or if it genuinely extracts knowledge from the source (e.g., Wikipedia biographies). In this paper, we conduct an in-depth study of this problem using a controlled set of semi-synthetic biography data. We uncover a relationship between the model's knowledge extraction ability and different diversity measures of the training data. We conduct (nearly) linear probing, revealing a strong correlation between this relationship and whether the model (nearly) linearly encodes the ...
In recent decades, the society depends more and more on computers for a large number of tasks. The f...
Because higher level cognitive processes generally involve the use of world knowledge, computational...
In this paper, we tackle the significant challenge of temporal knowledge reasoning in Large Language...
Recent progress in pretraining language models on large textual corpora led to a surge of improvemen...
Large language models show human-like performance in knowledge extraction, reasoning and dialogue, b...
Recent advances in large-scale pre-training provide large models with the potential to learn knowled...
Teaching new information to pre-trained large language models (PLM) is a crucial but challenging tas...
Neural language models have drastically changed the landscape of natural language processing (NLP). ...
Natural language has long been the most prominent tool for humans to disseminate, learn and create k...
Despite advances in deep learning and knowledge graphs (KGs), using language models for natural lang...
Thesis (Ph. D.)--University of Rochester. Dept. of Computer Science and Dept. of Linguistics, 2010.T...
Knowledge-intensive tasks, such as open-domain question answering (QA), require access to a large am...
While there has been tremendous progress in automatic database population in recent years, most of h...
Despite their wide adoption, the underlying training and memorization dynamics of very large languag...
The fluency and creativity of large pre-trained language models (LLMs) have led to their widespread ...
In recent decades, the society depends more and more on computers for a large number of tasks. The f...
Because higher level cognitive processes generally involve the use of world knowledge, computational...
In this paper, we tackle the significant challenge of temporal knowledge reasoning in Large Language...
Recent progress in pretraining language models on large textual corpora led to a surge of improvemen...
Large language models show human-like performance in knowledge extraction, reasoning and dialogue, b...
Recent advances in large-scale pre-training provide large models with the potential to learn knowled...
Teaching new information to pre-trained large language models (PLM) is a crucial but challenging tas...
Neural language models have drastically changed the landscape of natural language processing (NLP). ...
Natural language has long been the most prominent tool for humans to disseminate, learn and create k...
Despite advances in deep learning and knowledge graphs (KGs), using language models for natural lang...
Thesis (Ph. D.)--University of Rochester. Dept. of Computer Science and Dept. of Linguistics, 2010.T...
Knowledge-intensive tasks, such as open-domain question answering (QA), require access to a large am...
While there has been tremendous progress in automatic database population in recent years, most of h...
Despite their wide adoption, the underlying training and memorization dynamics of very large languag...
The fluency and creativity of large pre-trained language models (LLMs) have led to their widespread ...
In recent decades, the society depends more and more on computers for a large number of tasks. The f...
Because higher level cognitive processes generally involve the use of world knowledge, computational...
In this paper, we tackle the significant challenge of temporal knowledge reasoning in Large Language...