Understanding the effect of depth in deep learning is a critical problem. In this work, we utilize the Fourier analysis to empirically provide a promising mechanism to understand why feedforward deeper learning is faster. To this end, we separate a deep neural network, trained by normal stochastic gradient descent, into two parts during analysis, i.e., a pre-condition component and a learning component, in which the output of the pre-condition one is the input of the learning one. We use a filtering method to characterize the frequency distribution of a high-dimensional function. Based on experiments of deep networks and real dataset, we propose a deep frequency principle, that is, the effective target function for a deeper hidden layer...
Several decades ago, traditional neural networks were the most efficient machine learning technique....
Modern deep neural networks are highly over-parameterized compared to the data on which they are tra...
The weight initialization and the activation function of deep neural networks have a crucial impact ...
Understanding deep learning is increasingly emergent as it penetrates more and more into industry an...
© 2017, Springer Science+Business Media, LLC. We show how the success of deep learning could depend ...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
The paper reviews and extends an emerging body of theoretical results on deep learning including the...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
Despite the widespread practical success of deep learning methods, our theoretical understanding of ...
While the universal approximation property holds both for hierarchical and shallow networks, deep ne...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
This thesis investigates how general the knowledge stored in deep-Q-networks are. This general knowl...
© 2016 World Scientific Publishing Company. The paper briefly reviews several recent results on hier...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
Several decades ago, traditional neural networks were the most efficient machine learning technique....
Modern deep neural networks are highly over-parameterized compared to the data on which they are tra...
The weight initialization and the activation function of deep neural networks have a crucial impact ...
Understanding deep learning is increasingly emergent as it penetrates more and more into industry an...
© 2017, Springer Science+Business Media, LLC. We show how the success of deep learning could depend ...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
The paper characterizes classes of functions for which deep learning can be exponentially better tha...
The paper reviews and extends an emerging body of theoretical results on deep learning including the...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
Despite the widespread practical success of deep learning methods, our theoretical understanding of ...
While the universal approximation property holds both for hierarchical and shallow networks, deep ne...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
This thesis investigates how general the knowledge stored in deep-Q-networks are. This general knowl...
© 2016 World Scientific Publishing Company. The paper briefly reviews several recent results on hier...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
Several decades ago, traditional neural networks were the most efficient machine learning technique....
Modern deep neural networks are highly over-parameterized compared to the data on which they are tra...
The weight initialization and the activation function of deep neural networks have a crucial impact ...