Reasoning over natural language is a long-standing goal for the research community. However, studies have shown that existing language models are inadequate in reasoning. To address the issue, we present POET, a new pre-training paradigm. Through pre-training language models with programs and their execution results, POET empowers language models to harvest the reasoning knowledge possessed in program executors via a data-driven approach. POET is conceptually simple and can be instantiated by different kinds of programs. In this paper, we show three empirically powerful instances, i.e., POET-Math, POET-Logic, and POET-SQL. Experimental results on six benchmarks demonstrate that POET can significantly boost model performance on natural langu...
Inductive program synthesis, or inferring programs from examples of desired behavior, offers a gener...
The enormous rise in the scale, scope, and complexity of software projects has created a thriving ma...
Natural Language Inference (NLI) is considered a representative task to test natural language unders...
Language models have achieved remarkable performance on a wide range of tasks that require natural l...
Thesis (Ph. D.)--University of Rochester. Department of Computer Science, 2018.This dissertation exp...
Theorem proving in natural mathematical language - the mixture of symbolic and natural language used...
Despite recent success in large language model (LLM) reasoning, LLMs struggle with hierarchical mult...
Tackling Natural Language Inference with a logic-based method is becoming less and less common. Whil...
Humans understand language by extracting information (meaning) from sentences, combining it with exi...
Human language offers a powerful window into our thoughts -- we tell stories, give explanations, and...
Large-scale pre-trained language models (PLMs) bring new opportunities to challenge problems, especi...
Question-answering datasets require a broad set of reasoning skills. We show how to use question dec...
Recent Language Models (LMs) achieve breakthrough performance in code generation when trained on hum...
Neural-symbolic methods have shown their effectiveness in enhancing the reasoning abilities of large...
Through their transfer learning abilities, highly-parameterized large pre-trained language models ha...
Inductive program synthesis, or inferring programs from examples of desired behavior, offers a gener...
The enormous rise in the scale, scope, and complexity of software projects has created a thriving ma...
Natural Language Inference (NLI) is considered a representative task to test natural language unders...
Language models have achieved remarkable performance on a wide range of tasks that require natural l...
Thesis (Ph. D.)--University of Rochester. Department of Computer Science, 2018.This dissertation exp...
Theorem proving in natural mathematical language - the mixture of symbolic and natural language used...
Despite recent success in large language model (LLM) reasoning, LLMs struggle with hierarchical mult...
Tackling Natural Language Inference with a logic-based method is becoming less and less common. Whil...
Humans understand language by extracting information (meaning) from sentences, combining it with exi...
Human language offers a powerful window into our thoughts -- we tell stories, give explanations, and...
Large-scale pre-trained language models (PLMs) bring new opportunities to challenge problems, especi...
Question-answering datasets require a broad set of reasoning skills. We show how to use question dec...
Recent Language Models (LMs) achieve breakthrough performance in code generation when trained on hum...
Neural-symbolic methods have shown their effectiveness in enhancing the reasoning abilities of large...
Through their transfer learning abilities, highly-parameterized large pre-trained language models ha...
Inductive program synthesis, or inferring programs from examples of desired behavior, offers a gener...
The enormous rise in the scale, scope, and complexity of software projects has created a thriving ma...
Natural Language Inference (NLI) is considered a representative task to test natural language unders...