We can compare the expressiveness of neural networks that use rectified linear units (ReLUs) by the number of linear regions, which reflect the number of pieces of the piecewise linear functions modeled by such networks. However, enumerating these regions is prohibitive and the known analytical bounds are identical for networks with same dimensions. In this work, we approximate the number of linear regions through empirical bounds based on features of the trained network and probabilistic inference. Our first contribution is a method to sample the activation patterns defined by ReLUs using universal hash functions. This method is based on a Mixed-Integer Linear Programming (MILP) formulation of the network and an algorithm for probabilistic...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
In this article we present new results on neural networks with linear threshold activation functions...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
We can compare the expressiveness of neural networks that use rectified linear units (ReLUs) by the ...
Rectified linear units (ReLUs) have become the main model for the neural units in current deep learn...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
An important issue in neural network research is how to choose the number of nodes and layers such a...
We present results on the number of linear regions of the functions that can be represented by artif...
© 2019 Neural information processing systems foundation. All rights reserved. We study finite sample...
© 35th International Conference on Machine Learning, ICML 2018.All Rights Reserved. Verifying the r...
We algorithmically determine the regions and facets of all dimensions of the canonical polyhedral co...
We contribute to a better understanding of the class of functions that can be represented by a neura...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
Recently, path norm was proposed as a new capacity measure for neural networks with Rectified Linear...
One fundamental problem in deep learning is understanding the outstanding performance of deep Neural...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
In this article we present new results on neural networks with linear threshold activation functions...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
We can compare the expressiveness of neural networks that use rectified linear units (ReLUs) by the ...
Rectified linear units (ReLUs) have become the main model for the neural units in current deep learn...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
An important issue in neural network research is how to choose the number of nodes and layers such a...
We present results on the number of linear regions of the functions that can be represented by artif...
© 2019 Neural information processing systems foundation. All rights reserved. We study finite sample...
© 35th International Conference on Machine Learning, ICML 2018.All Rights Reserved. Verifying the r...
We algorithmically determine the regions and facets of all dimensions of the canonical polyhedral co...
We contribute to a better understanding of the class of functions that can be represented by a neura...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
Recently, path norm was proposed as a new capacity measure for neural networks with Rectified Linear...
One fundamental problem in deep learning is understanding the outstanding performance of deep Neural...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
In this article we present new results on neural networks with linear threshold activation functions...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...