In the past decade, neural networks have demonstrated impressive performance in supervised learning. They now power many applications ranging from real- time medical diagnosis to human-sounding virtual assistants through wild animal monitoring. Despite their increasing importance however, they remain difficult to train due to a complex interplay between the learning objective, the optimization algorithm and generalization performance. Indeed, using different loss functions and optimization algorithms lead to trained models with significantly different performances on unseen data. In this thesis, we focus first on the loss function, for which using a task-specific approach can improve the generalization performance in the small or noisy dat...
Deep neural networks have achieved significant success in a number of challenging engineering proble...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
Learning a deep neural network requires solving a challenging optimization problem: it is a high-dim...
Learning a deep neural network requires solving a challenging optimization problem: it is a high-dim...
The top-$k$ error is a common measure of performance in machine learning and computer vision. In pra...
How to train deep neural networks (DNNs) to generalize well is a central concern in deep learning, e...
The top-$k$ error is a common measure of performance in machine learning and computer vision. In pra...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
In modern supervised learning, many deep neural networks are able to interpolate the data: the empir...
Deep neural networks have achieved significant success in a number of challenging engineering proble...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
Learning a deep neural network requires solving a challenging optimization problem: it is a high-dim...
Learning a deep neural network requires solving a challenging optimization problem: it is a high-dim...
The top-$k$ error is a common measure of performance in machine learning and computer vision. In pra...
How to train deep neural networks (DNNs) to generalize well is a central concern in deep learning, e...
The top-$k$ error is a common measure of performance in machine learning and computer vision. In pra...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
In modern supervised learning, many deep neural networks are able to interpolate the data: the empir...
Deep neural networks have achieved significant success in a number of challenging engineering proble...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...
Learning in deep neural networks takes place by minimizing a nonconvex high-dimensional loss functio...