Deep neural networks (DNNs) for supervised learning can be viewed as a pipeline of the feature extractor (i.e., last hidden layer) and a linear classifier (i.e., output layer) that are trained jointly with stochastic gradient descent (SGD) on the loss function (e.g., cross-entropy). In each epoch, the true gradient of the loss function is estimated using a mini-batch sampled from the training set and model parameters are then updated with the mini-batch gradients. Although the latter provides an unbiased estimation of the former, they are subject to substantial variances derived from the size and number of sampled mini-batches, leading to noisy and jumpy updates. To stabilize such undesirable variance in estimating the true gradients, we pr...
Unsupervised joint alignment of images has been demonstrated to improve performance on recognition t...
While the backpropagation of error algorithm allowed for a rapid rise in the development and deploym...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...
Deep neural networks (DNNs) for supervised learning can be viewed as a pipeline of the feature extra...
Learning with noisy labels is one of the most practical but challenging tasks in deep learning. One ...
Direct feedback alignment (DFA) is emerging as an efficient and biologically plausible alternative t...
The performance of deep neural networks (DNNs) critically relies on high-quality annotations, while ...
Training with the true labels of a dataset as opposed to randomized labels leads to faster optimizat...
Inspired by two basic mechanisms in animal visual systems, we introduce a feature transform techniqu...
Adversarial examples easily mislead vision systems based on deep neural networks (DNNs) trained with...
Although deep neural networks have been proved effective in many applications, they are data hungry,...
Recent researches reveal that deep neural networks are sensitive to label noises hence leading to po...
This electronic version was submitted by the student author. The certified thesis is available in th...
Recent advances in large pre-trained models showed promising results in few-shot learning. However, ...
Despite being robust to small amounts of label noise, convolutional neural networks trained with sto...
Unsupervised joint alignment of images has been demonstrated to improve performance on recognition t...
While the backpropagation of error algorithm allowed for a rapid rise in the development and deploym...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...
Deep neural networks (DNNs) for supervised learning can be viewed as a pipeline of the feature extra...
Learning with noisy labels is one of the most practical but challenging tasks in deep learning. One ...
Direct feedback alignment (DFA) is emerging as an efficient and biologically plausible alternative t...
The performance of deep neural networks (DNNs) critically relies on high-quality annotations, while ...
Training with the true labels of a dataset as opposed to randomized labels leads to faster optimizat...
Inspired by two basic mechanisms in animal visual systems, we introduce a feature transform techniqu...
Adversarial examples easily mislead vision systems based on deep neural networks (DNNs) trained with...
Although deep neural networks have been proved effective in many applications, they are data hungry,...
Recent researches reveal that deep neural networks are sensitive to label noises hence leading to po...
This electronic version was submitted by the student author. The certified thesis is available in th...
Recent advances in large pre-trained models showed promising results in few-shot learning. However, ...
Despite being robust to small amounts of label noise, convolutional neural networks trained with sto...
Unsupervised joint alignment of images has been demonstrated to improve performance on recognition t...
While the backpropagation of error algorithm allowed for a rapid rise in the development and deploym...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...