We start from the contractive functional equation proposed in [4], where it was shown that the polynomial solution of functional equation can be used to initialize a Neural Network structure, with a controlled accuracy. We propose a novel algorithm, where the functional equation is solved with a converging iterative algorithm which can be realized as a Machine Learning training method iteratively with respect to the number of layers. The proof of convergence is performed with respect to the L ∞ norm. Numerical tests illustrate the theory and show that stochastic gradient descent methods can be used with good accuracy for this problem
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
We introduce a variational framework to learn the activation functions of deep neural networks. Our ...
This is Chapter 3 of the book titled "Deep Learning": a nine-part easy-to-grasp textbook written wit...
We consider deep neural networks, in which the output of each node is a quadratic function of its in...
Thesis: S.M., Massachusetts Institute of Technology, Sloan School of Management, Operations Research...
We study the computational complexity of (deterministic or randomized) algorithms based on point sam...
We present a multidimensional deep learning implementation of a stochastic branching algorithm for t...
Abstract. We prove that neural networks with a single hidden layer are capable of providing an optim...
The mathematical foundation of deep learning is the theorem that any continuous function can be appr...
We perform a comprehensive numerical study of the effect of approximation-theoretical results for ne...
The first part of this thesis develops fundamental limits of deep neural network learning by charact...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
An open problem around deep networks is the apparent absence of over-fitting despite large over-para...
Abstract. In this paper, a new algorithm for function approximation is proposed to obtain better gen...
We present a novel deep learning approach to approximate the solution of large, sparse, symmetric, p...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
We introduce a variational framework to learn the activation functions of deep neural networks. Our ...
This is Chapter 3 of the book titled "Deep Learning": a nine-part easy-to-grasp textbook written wit...
We consider deep neural networks, in which the output of each node is a quadratic function of its in...
Thesis: S.M., Massachusetts Institute of Technology, Sloan School of Management, Operations Research...
We study the computational complexity of (deterministic or randomized) algorithms based on point sam...
We present a multidimensional deep learning implementation of a stochastic branching algorithm for t...
Abstract. We prove that neural networks with a single hidden layer are capable of providing an optim...
The mathematical foundation of deep learning is the theorem that any continuous function can be appr...
We perform a comprehensive numerical study of the effect of approximation-theoretical results for ne...
The first part of this thesis develops fundamental limits of deep neural network learning by charact...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
An open problem around deep networks is the apparent absence of over-fitting despite large over-para...
Abstract. In this paper, a new algorithm for function approximation is proposed to obtain better gen...
We present a novel deep learning approach to approximate the solution of large, sparse, symmetric, p...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
We introduce a variational framework to learn the activation functions of deep neural networks. Our ...
This is Chapter 3 of the book titled "Deep Learning": a nine-part easy-to-grasp textbook written wit...