Characterizing the implicit structure of the computation within neural networks is a foundational problem in the area of deep learning interpretability. Can their inner decision process be captured symbolically in some familiar logic? We show that any transformer neural network can be translated into an equivalent fixed-size first-order logic formula which may also use majority quantifiers. The idea is to simulate transformers with highly uniform threshold circuits and leverage known theoretical connections between circuits and logic. Our findings also reveal the surprising fact that the entire transformer computation can be reduced merely to the division of two (large) integers. While our results are most pertinent for transformers, they a...
We present a novel method for scalable and precise certification of deep neural networks. The key te...
The transformer is a neural network component that can be used to learn useful representations of se...
Abstract. It is shown that high-order feedforward neural nets of constant depth with piecewise-polyn...
Characterizing neural networks in terms of better-understood formal systems has the potential to yie...
Studying symbolic computation in deep neural networks (DNNs) is essential for improving their explai...
Recent theoretical work has identified surprisingly simple reasoning problems, such as checking if t...
This document aims to be a self-contained, mathematically precise overview of transformer architectu...
We study two fundamental questions in neuro-symbolic computing: can deep learning tackle challenging...
We propose a synthetic task, LEGO (Learning Equality and Group Operations), that encapsulates the pr...
Transformer based models are used to achieve state-of-the-art performance on various deep learning t...
The integration of symbolic and neural-network-based artificial intelligence paradigms constitutes a...
The integration of symbolic and neural-network-based artificial intelligence paradigms constitutes ...
We show how to "compile" human-readable programs into standard decoder-only transformer models. Our ...
Mathematical reasoning is one of the most impressive achievements of human intellect but remains a f...
Transformer networks have seen great success in natural language processing and machine vision, wher...
We present a novel method for scalable and precise certification of deep neural networks. The key te...
The transformer is a neural network component that can be used to learn useful representations of se...
Abstract. It is shown that high-order feedforward neural nets of constant depth with piecewise-polyn...
Characterizing neural networks in terms of better-understood formal systems has the potential to yie...
Studying symbolic computation in deep neural networks (DNNs) is essential for improving their explai...
Recent theoretical work has identified surprisingly simple reasoning problems, such as checking if t...
This document aims to be a self-contained, mathematically precise overview of transformer architectu...
We study two fundamental questions in neuro-symbolic computing: can deep learning tackle challenging...
We propose a synthetic task, LEGO (Learning Equality and Group Operations), that encapsulates the pr...
Transformer based models are used to achieve state-of-the-art performance on various deep learning t...
The integration of symbolic and neural-network-based artificial intelligence paradigms constitutes a...
The integration of symbolic and neural-network-based artificial intelligence paradigms constitutes ...
We show how to "compile" human-readable programs into standard decoder-only transformer models. Our ...
Mathematical reasoning is one of the most impressive achievements of human intellect but remains a f...
Transformer networks have seen great success in natural language processing and machine vision, wher...
We present a novel method for scalable and precise certification of deep neural networks. The key te...
The transformer is a neural network component that can be used to learn useful representations of se...
Abstract. It is shown that high-order feedforward neural nets of constant depth with piecewise-polyn...