International audienceDuring training, the weights of a Deep Neural Network (DNN) are optimized from a random initialization towards a nearly optimum value minimizing a loss function. Only this final state of the weights is typically kept for testing, while the wealth of information on the geometry of the weight space, accumulated over the descent towards the minimum is discarded. In this work we propose to make use of this knowledge and leverage it for computing the distributions of the weights of the DNN. This can be further used for estimating the epistemic uncertainty of the DNN by aggregating predictions from an ensemble of networks sampled from these distributions. To this end we introduce a method for tracking the trajectory of the w...
We present an approach to quantifying both aleatoric and epistemic uncertainty for deep neural netwo...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
We present a new type of probabilistic model which we call DISsimilarity COefficient Networks (DISCO...
Artificial Intelligent (AI) has become the most potent and forward-looking force in the technologies...
Supervised training of deep neural nets typically relies on minimizing cross-entropy. However, in ma...
Deep neural networks are powerful predictors for a variety of tasks. However, they do not capture un...
The performance of deep learning (DL) models is highly dependent on the quality and size of the trai...
International audienceIn the context of supervised learning of a function by a neural network, we cl...
A new method of initializing the weights in deep neural networks is proposed. The method follows two...
Stochastic variational inference for Bayesian deep neural network (DNN) requires specifying priors a...
Deep Neural Networks ("deep learning") have become a ubiquitous choice of algorithms for Machine Lea...
This paper considers the generation of prediction intervals (PIs) by neural networks for quantifying...
Deep learning uses neural networks which are parameterised by their weights. The neural networks ar...
We study the dynamics of gradient descent in learning neural networks for classification problems. U...
International audienceDeep neural networks (DNNs) are powerful learning models yet their results are...
We present an approach to quantifying both aleatoric and epistemic uncertainty for deep neural netwo...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
We present a new type of probabilistic model which we call DISsimilarity COefficient Networks (DISCO...
Artificial Intelligent (AI) has become the most potent and forward-looking force in the technologies...
Supervised training of deep neural nets typically relies on minimizing cross-entropy. However, in ma...
Deep neural networks are powerful predictors for a variety of tasks. However, they do not capture un...
The performance of deep learning (DL) models is highly dependent on the quality and size of the trai...
International audienceIn the context of supervised learning of a function by a neural network, we cl...
A new method of initializing the weights in deep neural networks is proposed. The method follows two...
Stochastic variational inference for Bayesian deep neural network (DNN) requires specifying priors a...
Deep Neural Networks ("deep learning") have become a ubiquitous choice of algorithms for Machine Lea...
This paper considers the generation of prediction intervals (PIs) by neural networks for quantifying...
Deep learning uses neural networks which are parameterised by their weights. The neural networks ar...
We study the dynamics of gradient descent in learning neural networks for classification problems. U...
International audienceDeep neural networks (DNNs) are powerful learning models yet their results are...
We present an approach to quantifying both aleatoric and epistemic uncertainty for deep neural netwo...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
We present a new type of probabilistic model which we call DISsimilarity COefficient Networks (DISCO...