Modern neural networks often have great expressive power and can be trained to overfit the training data, while still achieving a good test performance. This phenomenon is referred to as "benign overfitting". Recently, there emerges a line of works studying "benign overfitting" from the theoretical perspective. However, they are limited to linear models or kernel/random feature models, and there is still a lack of theoretical understanding about when and how benign overfitting occurs in neural networks. In this paper, we study the benign overfitting phenomenon in training a two-layer convolutional neural network (CNN). We show that when the signal-to-noise ratio satisfies a certain condition, a two-layer CNN trained by gradient descent can ...
Neural networks are known to be highly sensitive to adversarial examples. These may arise due to dif...
CNN Filter DB: An Empirical Investigation of Trained Convolutional. Poster as presented at CVPR2022...
The risk of overparameterized models, in particular deep neural networks, is often double-descent sh...
Benign overfitting, the phenomenon where interpolating models generalize well in the presence of noi...
Modern machine learning often operates in the regime where the number of parameters is much higher t...
The recent success of neural network models has shone light on a rather surprising statistical pheno...
We investigate two causes for adversarial vulnerability in deep neural networks: bad data and (poorl...
Traditional convolutional neural networks exhibit an inherent limitation, they can not adapt their c...
When a large feedforward neural network is trained on a small training set, it typically performs po...
Current deep neural networks are highly overparameterized (up to billions of connection weights) and...
In this paper we examine a perceptron learning task. The task is realizable since it is provided by ...
Some applications have the property of being resilient, meaning that they are robust to noise (e.g. ...
For many reasons, neural networks have become very popular AI machine learning models. Two of the mo...
The remarkable practical success of deep learning has revealed some major surprises from a theoretic...
In recent years, deep neural networks have become the state-of-the art in many machine learning doma...
Neural networks are known to be highly sensitive to adversarial examples. These may arise due to dif...
CNN Filter DB: An Empirical Investigation of Trained Convolutional. Poster as presented at CVPR2022...
The risk of overparameterized models, in particular deep neural networks, is often double-descent sh...
Benign overfitting, the phenomenon where interpolating models generalize well in the presence of noi...
Modern machine learning often operates in the regime where the number of parameters is much higher t...
The recent success of neural network models has shone light on a rather surprising statistical pheno...
We investigate two causes for adversarial vulnerability in deep neural networks: bad data and (poorl...
Traditional convolutional neural networks exhibit an inherent limitation, they can not adapt their c...
When a large feedforward neural network is trained on a small training set, it typically performs po...
Current deep neural networks are highly overparameterized (up to billions of connection weights) and...
In this paper we examine a perceptron learning task. The task is realizable since it is provided by ...
Some applications have the property of being resilient, meaning that they are robust to noise (e.g. ...
For many reasons, neural networks have become very popular AI machine learning models. Two of the mo...
The remarkable practical success of deep learning has revealed some major surprises from a theoretic...
In recent years, deep neural networks have become the state-of-the art in many machine learning doma...
Neural networks are known to be highly sensitive to adversarial examples. These may arise due to dif...
CNN Filter DB: An Empirical Investigation of Trained Convolutional. Poster as presented at CVPR2022...
The risk of overparameterized models, in particular deep neural networks, is often double-descent sh...