International audienceDeep learning has been immensely successful at a variety of tasks, ranging from classification to artificial intelligence. Learning corresponds to fitting training data, which is implemented by descending a very high-dimensional loss function. Understanding under which conditions neural networks do not get stuck in poor minima of the loss, and how the landscape of that loss evolves as depth is increased, remains a challenge. Here we predict, and test empirically, an analogy between this landscape and the energy landscape of repulsive ellipses. We argue that in fully connected deep networks a phase transition delimits the over- and underparametrized regimes where fitting can or cannot be achieved. In the vicinity of thi...
The success of deep learning has revealed the application potential of neural networks across the sc...
Loss landscape analysis is extremely useful for a deeper understanding of the generalization ability...
Loss landscape analysis is extremely useful for a deeper understanding of the generalization ability...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
Deep learning has been immensely successful at a variety of tasks, ranging from classification to ar...
Current deep neural networks are highly overparameterized (up to billions of connection weights) and...
Gradient descent dynamics in complex energy landscapes, i.e. featuring multiple minima, finds applic...
We explore the loss landscape of fully-connected and convolutional neural networks using random, low...
peer reviewedGradient descent dynamics in complex energy landscapes, i.e. featuring multiple minima,...
Abstract: We investigate the structure of the loss function landscape for neural networks subject to...
The success of deep learning has revealed the application potential of neural networks across the sc...
Abstract: We investigate the structure of the loss function landscape for neural networks subject to...
The success of deep learning has revealed the application potential of neural networks across the sc...
Loss landscape analysis is extremely useful for a deeper understanding of the generalization ability...
Loss landscape analysis is extremely useful for a deeper understanding of the generalization ability...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
International audienceDeep learning has been immensely successful at a variety of tasks, ranging fro...
Deep learning has been immensely successful at a variety of tasks, ranging from classification to ar...
Current deep neural networks are highly overparameterized (up to billions of connection weights) and...
Gradient descent dynamics in complex energy landscapes, i.e. featuring multiple minima, finds applic...
We explore the loss landscape of fully-connected and convolutional neural networks using random, low...
peer reviewedGradient descent dynamics in complex energy landscapes, i.e. featuring multiple minima,...
Abstract: We investigate the structure of the loss function landscape for neural networks subject to...
The success of deep learning has revealed the application potential of neural networks across the sc...
Abstract: We investigate the structure of the loss function landscape for neural networks subject to...
The success of deep learning has revealed the application potential of neural networks across the sc...
Loss landscape analysis is extremely useful for a deeper understanding of the generalization ability...
Loss landscape analysis is extremely useful for a deeper understanding of the generalization ability...