The top-k classification accuracy is one of the core metrics in machine learning. Here, k is conventionally a positive integer, such as 1 or 5, leading to top-1 or top-5 training objectives. In this work, we relax this assumption and optimize the model for multiple k simultaneously instead of using a single k. Leveraging recent advances in differentiable sorting and ranking, we propose a differentiable top-k cross-entropy classification loss. This allows training the network while not only considering the top-1 prediction, but also, e.g., the top-2 and top-5 predictions. We evaluate the proposed loss function for fine-tuning on state-of-the-art architectures, as well as for training from scratch. We find that relaxing k does not only produc...
When humans learn a new concept, they might ignore examples that they cannot make sense of at first,...
Mutual Information (MI) has been widely used as a loss regularizer for training neural networks. Thi...
Bipartite ranking aims to learn a real-valued ranking function that orders positive instances before...
The top-$k$ error is a common measure of performance in machine learning and computer vision. In pra...
In the past decade, neural networks have demonstrated impressive performance in supervised learning....
International audienceIn modern classification tasks, the number of labels is getting larger and lar...
Average-K classification is an alternative to top-K classification in which the number of labels ret...
ListNet is a well-known listwise learning to rank model and has gained much atten-tion in recent yea...
In the process of machine learning, models are essentially defined by a group of parameters in multi...
A large amount of research on Convolutional Neural Networks (CNN) has focused on flat Classification...
International audienceThe problem of ranking a set of visual samples according to their relevance to...
Is it possible to train several classifiers to perform meaningful crowd-sourcing to produce a better...
Deep learning has been shown to achieve impressive results in several domains like computer vision a...
Over the past decade, deep neural networks have proven to be adept in image classification tasks, of...
This paper studies probability distributions of penultimate activations of classification networks. ...
When humans learn a new concept, they might ignore examples that they cannot make sense of at first,...
Mutual Information (MI) has been widely used as a loss regularizer for training neural networks. Thi...
Bipartite ranking aims to learn a real-valued ranking function that orders positive instances before...
The top-$k$ error is a common measure of performance in machine learning and computer vision. In pra...
In the past decade, neural networks have demonstrated impressive performance in supervised learning....
International audienceIn modern classification tasks, the number of labels is getting larger and lar...
Average-K classification is an alternative to top-K classification in which the number of labels ret...
ListNet is a well-known listwise learning to rank model and has gained much atten-tion in recent yea...
In the process of machine learning, models are essentially defined by a group of parameters in multi...
A large amount of research on Convolutional Neural Networks (CNN) has focused on flat Classification...
International audienceThe problem of ranking a set of visual samples according to their relevance to...
Is it possible to train several classifiers to perform meaningful crowd-sourcing to produce a better...
Deep learning has been shown to achieve impressive results in several domains like computer vision a...
Over the past decade, deep neural networks have proven to be adept in image classification tasks, of...
This paper studies probability distributions of penultimate activations of classification networks. ...
When humans learn a new concept, they might ignore examples that they cannot make sense of at first,...
Mutual Information (MI) has been widely used as a loss regularizer for training neural networks. Thi...
Bipartite ranking aims to learn a real-valued ranking function that orders positive instances before...