Consider the multivariate nonparametric regression model. It is shown that estimators based on sparsely connected deep neural networks with ReLU activation function and properly chosen network architecture achieve the minimax rates of convergence (up to logn logn -factors) under a general composition assumption on the regression function. The framework includes many well-studied structural constraints such as (generalized) additive models. While there is a lot of flexibility in the network architecture, the tuning parameter is the sparsity of the network. Specifically, we consider large networks with number of potential network parameters exceeding the sample size. The analysis gives some insights into why multilayer feedforward neural net...
N-tuple neural networks (NTNNs) have been successfully applied to both pattern recognition and funct...
We explore convergence of deep neural networks with the popular ReLU activation function, as the dep...
Abstract-This paper describes a memory-based network that provides estimates of continuous variables...
Consider the multivariate nonparametric regression model. It is shown that estimators based on spars...
We study the theory of neural network (NN) from the lens of classical nonparametric regression probl...
It is a central problem in both statistics and computer science to understand the theoretical founda...
In theory, recent results in nonparametric regression show that neural network estimates are able to...
In recent years, modern technology has facilitated the collection of large-scale data from medical r...
Models built with deep neural network (DNN) can handle complicated real-world data extremely well, s...
In this paper, we consider estimation and inference for both the multi-index parameters and the link...
Deep neural networks (DNNs) generate much richer function spaces than shallow networks. Since the fu...
In this paper, a new nonsmooth optimization based algorithm for solving large-scale regression probl...
Deep neural networks, as a powerful system to represent high dimensional complex functions, play a k...
Deep neural networks are the main subject of interest in the study of theoretical deep learning, whi...
Deep feedforward neural networks with piecewise linear activations are currently producing the state...
N-tuple neural networks (NTNNs) have been successfully applied to both pattern recognition and funct...
We explore convergence of deep neural networks with the popular ReLU activation function, as the dep...
Abstract-This paper describes a memory-based network that provides estimates of continuous variables...
Consider the multivariate nonparametric regression model. It is shown that estimators based on spars...
We study the theory of neural network (NN) from the lens of classical nonparametric regression probl...
It is a central problem in both statistics and computer science to understand the theoretical founda...
In theory, recent results in nonparametric regression show that neural network estimates are able to...
In recent years, modern technology has facilitated the collection of large-scale data from medical r...
Models built with deep neural network (DNN) can handle complicated real-world data extremely well, s...
In this paper, we consider estimation and inference for both the multi-index parameters and the link...
Deep neural networks (DNNs) generate much richer function spaces than shallow networks. Since the fu...
In this paper, a new nonsmooth optimization based algorithm for solving large-scale regression probl...
Deep neural networks, as a powerful system to represent high dimensional complex functions, play a k...
Deep neural networks are the main subject of interest in the study of theoretical deep learning, whi...
Deep feedforward neural networks with piecewise linear activations are currently producing the state...
N-tuple neural networks (NTNNs) have been successfully applied to both pattern recognition and funct...
We explore convergence of deep neural networks with the popular ReLU activation function, as the dep...
Abstract-This paper describes a memory-based network that provides estimates of continuous variables...