While non-linear activation functions play vital roles in artificial neural networks, it is generally unclear how the non-linearity can improve the quality of function approximations. In this paper, we present a theoretical framework to rigorously analyze the performance gain of using non-linear activation functions for a class of residual neural networks (ResNets). In particular, we show that when the input features for the ResNet are uniformly chosen and orthogonal to each other, using non-linear activation functions to generate the ResNet output averagely outperforms using linear activation functions, and the performance gain can be explicitly computed. Moreover, we show that when the activation functions are chosen as polynomials with t...
In this paper, we study the theoretical properties of a new kind of artificial neural network, which...
The multiplicity of approximation theorems for Neural Networks do not relate to approximation of lin...
In the paper, an ontogenic artificial neural network (ANNs) is proposed. The network uses orthogonal...
We consider neural networks with rational activation functions. The choice of the nonlinear activati...
Artificial neural networks are function-approximating models that can improve themselves with experi...
Abstract. We prove that neural networks with a single hidden layer are capable of providing an optim...
In this article we present new results on neural networks with linear threshold activation functions...
The activation function deployed in a deep neural network has great influence on the performance of ...
© 2018 Curran Associates Inc..All rights reserved. Finding minimum distortion of adversarial example...
We compare activation functions in terms of the approximation power of their feedforward nets. We co...
Neural networks are generally built by interleaving (adaptable) linear layers with (fixed) nonlinear...
We introduce a variational framework to learn the activation functions of deep neural networks. Our ...
An activation function, possibly new, is proposed for use in digital simulation of arti#cial neural ...
This report introduces a novel algorithm to learn the width of non-linear activation functions (of a...
Abstract—A technique for approximating a continuous function of variables with a radial basis functi...
In this paper, we study the theoretical properties of a new kind of artificial neural network, which...
The multiplicity of approximation theorems for Neural Networks do not relate to approximation of lin...
In the paper, an ontogenic artificial neural network (ANNs) is proposed. The network uses orthogonal...
We consider neural networks with rational activation functions. The choice of the nonlinear activati...
Artificial neural networks are function-approximating models that can improve themselves with experi...
Abstract. We prove that neural networks with a single hidden layer are capable of providing an optim...
In this article we present new results on neural networks with linear threshold activation functions...
The activation function deployed in a deep neural network has great influence on the performance of ...
© 2018 Curran Associates Inc..All rights reserved. Finding minimum distortion of adversarial example...
We compare activation functions in terms of the approximation power of their feedforward nets. We co...
Neural networks are generally built by interleaving (adaptable) linear layers with (fixed) nonlinear...
We introduce a variational framework to learn the activation functions of deep neural networks. Our ...
An activation function, possibly new, is proposed for use in digital simulation of arti#cial neural ...
This report introduces a novel algorithm to learn the width of non-linear activation functions (of a...
Abstract—A technique for approximating a continuous function of variables with a radial basis functi...
In this paper, we study the theoretical properties of a new kind of artificial neural network, which...
The multiplicity of approximation theorems for Neural Networks do not relate to approximation of lin...
In the paper, an ontogenic artificial neural network (ANNs) is proposed. The network uses orthogonal...