We contribute to a better understanding of the class of functions that can be represented by a neural network with ReLU activations and a given architecture. Using techniques from mixed-integer optimization, polyhedral theory, and tropical geometry, we provide a mathematical counterbalance to the universal approximation theorems which suggest that a single hidden layer is sufficient for learning any function. In particular, we investigate whether the class of exactly representable functions strictly increases by adding more layers (with no restrictions on size). As a by-product of our investigations, we settle an old conjecture about piecewise linear functions by Wang and Sun [IEEE Trans. Inform. Theory, 51 (2005), pp. 4425-4431] in the aff...
This paper develops simple feed-forward neural networks that achieve the universal approximation pro...
This is Chapter 2 of Part 1 of the book titled "Deep Learning": a nine-part easy-to-grasp textbook w...
We study the expressive power of deep ReLU neural networks for approximating functions in dilated sh...
We contribute to a better understanding of the class of functions that can be represented by a neura...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
We study the necessary and sufficient complexity of ReLU neural networks---in terms of depth and num...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
The paper briefly reviews several recent results on hierarchical architectures for learning from exa...
We solve an open question from Lu et al. (2017), by showing that any target network with inputs in $...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
We establish in this work approximation results of deep neural networks for smooth functions measure...
International audienceWe study the expressivity of deep neural networks. Measuring a network's compl...
The first part of this thesis develops fundamental limits of deep neural network learning by charact...
AbstractApproximation properties of the MLP (multilayer feedforward perceptron) model of neural netw...
This paper develops simple feed-forward neural networks that achieve the universal approximation pro...
This is Chapter 2 of Part 1 of the book titled "Deep Learning": a nine-part easy-to-grasp textbook w...
We study the expressive power of deep ReLU neural networks for approximating functions in dilated sh...
We contribute to a better understanding of the class of functions that can be represented by a neura...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
We study the necessary and sufficient complexity of ReLU neural networks---in terms of depth and num...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
The paper briefly reviews several recent results on hierarchical architectures for learning from exa...
We solve an open question from Lu et al. (2017), by showing that any target network with inputs in $...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
We establish in this work approximation results of deep neural networks for smooth functions measure...
International audienceWe study the expressivity of deep neural networks. Measuring a network's compl...
The first part of this thesis develops fundamental limits of deep neural network learning by charact...
AbstractApproximation properties of the MLP (multilayer feedforward perceptron) model of neural netw...
This paper develops simple feed-forward neural networks that achieve the universal approximation pro...
This is Chapter 2 of Part 1 of the book titled "Deep Learning": a nine-part easy-to-grasp textbook w...
We study the expressive power of deep ReLU neural networks for approximating functions in dilated sh...