While the universal approximation property holds both for hierarchical and shallow networks, deep networks can approximate the class of compositional functions as well as shallow networks but with exponentially lower number of training parameters and sample complexity. Compositional functions are obtained as a hierarchy of local constituent functions, where "local functions'' are functions with low dimensionality. This theorem proves an old conjecture by Bengio on the role of depth in networks, characterizing precisely the conditions under which it holds. It also suggests possible answers to the the puzzle of why high-dimensional deep networks trained on large training sets often do not seem to show overfit
Recently, deep networks were proved to be more effective than shallow architectures to face complex ...
National audienceWe study the expressivity of sparsely connected deep networks. Measuring a network'...
Modern deep neural networks are highly over-parameterized compared to the data on which they are tra...
© 2020 American Institute of Mathematical Sciences. All rights reserved. We show that deep networks ...
Deep learning networks with convolution, pooling and subsampling are a special case of hierar- chica...
We describe computational tasks - especially in vision - that correspond to compositional/hierarchic...
The paper reviews and extends an emerging body of theoretical results on deep learning including the...
The paper briefly reviews several recent results on hierarchical architectures for learning from exa...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
We investigate the representational power of sum-product networks (computation networks analogous to...
© 2020 National Academy of Sciences. All rights reserved. While deep learning is successful in a num...
The main success stories of deep learning, starting with ImageNet, depend on convolutional networks,...
Recently, deep networks were proved to be more effective than shallow architectures to face complex ...
National audienceWe study the expressivity of sparsely connected deep networks. Measuring a network'...
Modern deep neural networks are highly over-parameterized compared to the data on which they are tra...
© 2020 American Institute of Mathematical Sciences. All rights reserved. We show that deep networks ...
Deep learning networks with convolution, pooling and subsampling are a special case of hierar- chica...
We describe computational tasks - especially in vision - that correspond to compositional/hierarchic...
The paper reviews and extends an emerging body of theoretical results on deep learning including the...
The paper briefly reviews several recent results on hierarchical architectures for learning from exa...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
We investigate the representational power of sum-product networks (computation networks analogous to...
© 2020 National Academy of Sciences. All rights reserved. While deep learning is successful in a num...
The main success stories of deep learning, starting with ImageNet, depend on convolutional networks,...
Recently, deep networks were proved to be more effective than shallow architectures to face complex ...
National audienceWe study the expressivity of sparsely connected deep networks. Measuring a network'...
Modern deep neural networks are highly over-parameterized compared to the data on which they are tra...