Reliable generalization lies at the heart of safe ML and AI. However, understanding when and how neural networks generalize remains one of the most important unsolved problems in the field. In this work, we conduct an extensive empirical study (2200 models, 16 tasks) to investigate whether insights from the theory of computation can predict the limits of neural network generalization in practice. We demonstrate that grouping tasks according to the Chomsky hierarchy allows us to forecast whether certain architectures will be able to generalize to out-of-distribution inputs. This includes negative results where even extensive amounts of data and training time never led to any non-trivial generalization, despite models having sufficient capaci...
Deep learning has transformed computer vision, natural language processing, and speech recognition. ...
A key feature of intelligent behaviour is the ability to learn abstract strategies that scale and tr...
The power of human language and thought arises from systematic compositionality—the algebraic abilit...
This thesis presents a new theory of generalization in neural network types of learning machines. Th...
With a direct analysis of neural networks, this paper presents a mathematically tight generalization...
By making assumptions on the probability distribution of the potentials in a feed-forward neural net...
We present a unified framework for a number of different ways of failing to generalize properly. Du...
We present a unified framework for a number of different ways of failing to generalize properly. Dur...
Over-paramaterized neural models have become dominant in Natural Language Processing. Increasing the...
Generalization is a central aspect of learning theory. Here, we propose a framework that explores an...
For decades research has pursued the ambitious goal of designing computer models that learn to solve...
\Ve describe a series of careful llumerical experiments which measure the average generalization cap...
Artificial neural networks have become highly effective at performing specific, challenging tasks by...
Usually, generalization is considered as a function of learning from a set of examples. In present w...
Recurrent Neural Networks (RNNs) are theoretically Turing-complete and established themselves as a d...
Deep learning has transformed computer vision, natural language processing, and speech recognition. ...
A key feature of intelligent behaviour is the ability to learn abstract strategies that scale and tr...
The power of human language and thought arises from systematic compositionality—the algebraic abilit...
This thesis presents a new theory of generalization in neural network types of learning machines. Th...
With a direct analysis of neural networks, this paper presents a mathematically tight generalization...
By making assumptions on the probability distribution of the potentials in a feed-forward neural net...
We present a unified framework for a number of different ways of failing to generalize properly. Du...
We present a unified framework for a number of different ways of failing to generalize properly. Dur...
Over-paramaterized neural models have become dominant in Natural Language Processing. Increasing the...
Generalization is a central aspect of learning theory. Here, we propose a framework that explores an...
For decades research has pursued the ambitious goal of designing computer models that learn to solve...
\Ve describe a series of careful llumerical experiments which measure the average generalization cap...
Artificial neural networks have become highly effective at performing specific, challenging tasks by...
Usually, generalization is considered as a function of learning from a set of examples. In present w...
Recurrent Neural Networks (RNNs) are theoretically Turing-complete and established themselves as a d...
Deep learning has transformed computer vision, natural language processing, and speech recognition. ...
A key feature of intelligent behaviour is the ability to learn abstract strategies that scale and tr...
The power of human language and thought arises from systematic compositionality—the algebraic abilit...