Investigating the reasoning abilities of transformer models, and discovering new challenging tasks for them, has been a topic of much interest. Recent studies have found these models to be surprisingly strong at performing deductive reasoning over formal logical theories expressed in natural language. A shortcoming of these studies, however, is that they do not take into account that logical theories, when sampled uniformly at random, do not necessarily lead to hard instances. We propose a new methodology for creating challenging algorithmic reasoning datasets that focus on natural language satisfiability (NLSat) problems. The key idea is to draw insights from empirical sampling of hard propositional SAT problems and from complexity-theoret...
Deductive reasoning (drawing conclusions from assumptions) is a challenging problem in NLP. In this ...
Large language models (LLMs) have shown remarkable reasoning capabilities given chain-of-thought pro...
Logical reasoning is needed in a wide range of NLP tasks. Can a BERT model be trained end-to-end to ...
Transformers have been shown to be able to perform deductive reasoning on a logical rulebase contain...
One way that the current state of the art measures the reasoning ability of transformer-based models...
AbstractVery few natural language understanding applications employ methods from automated deduction...
Thesis (Ph. D.)--University of Rochester. Department of Computer Science, 2018.This dissertation exp...
Tackling Natural Language Inference with a logic-based method is becoming less and less common. Whil...
Reasoning with knowledge expressed in natural language and Knowledge Bases (KBs) is a major challen...
Strangely enough, it is possible to use machine learning models to predict the satisfiability status...
Research in automated deduction is traditionally focused on the problem of determining the satisfia...
Natural language understanding (NLU) of text is a fundamental challenge in AI, and it has received s...
While pre-trained language models (PLMs) are the go-to solution to tackle many natural language proc...
Large-scale, pre-trained language models (LMs) have achieved human-level performance on a breadth of...
We report results from large-scale experiments in satisfiability testing. As has been observed by ot...
Deductive reasoning (drawing conclusions from assumptions) is a challenging problem in NLP. In this ...
Large language models (LLMs) have shown remarkable reasoning capabilities given chain-of-thought pro...
Logical reasoning is needed in a wide range of NLP tasks. Can a BERT model be trained end-to-end to ...
Transformers have been shown to be able to perform deductive reasoning on a logical rulebase contain...
One way that the current state of the art measures the reasoning ability of transformer-based models...
AbstractVery few natural language understanding applications employ methods from automated deduction...
Thesis (Ph. D.)--University of Rochester. Department of Computer Science, 2018.This dissertation exp...
Tackling Natural Language Inference with a logic-based method is becoming less and less common. Whil...
Reasoning with knowledge expressed in natural language and Knowledge Bases (KBs) is a major challen...
Strangely enough, it is possible to use machine learning models to predict the satisfiability status...
Research in automated deduction is traditionally focused on the problem of determining the satisfia...
Natural language understanding (NLU) of text is a fundamental challenge in AI, and it has received s...
While pre-trained language models (PLMs) are the go-to solution to tackle many natural language proc...
Large-scale, pre-trained language models (LMs) have achieved human-level performance on a breadth of...
We report results from large-scale experiments in satisfiability testing. As has been observed by ot...
Deductive reasoning (drawing conclusions from assumptions) is a challenging problem in NLP. In this ...
Large language models (LLMs) have shown remarkable reasoning capabilities given chain-of-thought pro...
Logical reasoning is needed in a wide range of NLP tasks. Can a BERT model be trained end-to-end to ...