Transformers, currently the state-of-the-art in natural language understanding (NLU) tasks, are prone to generate uncalibrated predictions or extreme probabilities, making the process of taking different decisions based on their output relatively difficult. In this paper we propose to build several inductive Venn--ABERS predictors (IVAP), which are guaranteed to be well calibrated under minimal assumptions, based on a selection of pre-trained transformers. We test their performance over a set of diverse NLU tasks and show that they are capable of producing well-calibrated probabilistic predictions that are uniformly spread over the [0,1] interval -- all while retaining the original model's predictive accuracy.Comment: Accepted at the 11th S...
Deep neural networks are powerful tools to detect hidden patterns in data and leverage them to make ...
Successful use of probabilistic classification requires well-calibrated probability estimates, i.e.,...
Pretrained multilingual encoder models can directly perform zero-shot multilingual tasks or linguist...
Inductive Venn-ABERS predictors (IVAPs) are a type of probabilistic predictors with the theoretical ...
Inductive Venn-ABERS predictors (IVAPs) are a type of probabilistic predictors with the theoretical ...
We propose a new approach for constructing prediction sets for Transformer networks via the strong s...
AI tools can be useful to address model deficits in the design of communication systems. However, co...
Probabilistic classification requires well-calibrated probability estimates, i.e., the predicted cla...
The task of abductive natural language inference (\alpha{}nli), to decide which hypothesis is the mo...
With model trustworthiness being crucial for sensitive real-world applications, practitioners are pu...
Adding confidence measures to predictive models should increase the trustworthiness, but only if the...
Adding confidence measures to predictive models should increase the trustworthiness, but only if the...
Model calibration, which is concerned with how frequently the model predicts correctly, not only pla...
Part 4: First Conformal Prediction and Its Applications Workshop (COPA 2012)International audienceVe...
Accurate uncertainty quantification is necessary to enhance the reliability of deep learning models ...
Deep neural networks are powerful tools to detect hidden patterns in data and leverage them to make ...
Successful use of probabilistic classification requires well-calibrated probability estimates, i.e.,...
Pretrained multilingual encoder models can directly perform zero-shot multilingual tasks or linguist...
Inductive Venn-ABERS predictors (IVAPs) are a type of probabilistic predictors with the theoretical ...
Inductive Venn-ABERS predictors (IVAPs) are a type of probabilistic predictors with the theoretical ...
We propose a new approach for constructing prediction sets for Transformer networks via the strong s...
AI tools can be useful to address model deficits in the design of communication systems. However, co...
Probabilistic classification requires well-calibrated probability estimates, i.e., the predicted cla...
The task of abductive natural language inference (\alpha{}nli), to decide which hypothesis is the mo...
With model trustworthiness being crucial for sensitive real-world applications, practitioners are pu...
Adding confidence measures to predictive models should increase the trustworthiness, but only if the...
Adding confidence measures to predictive models should increase the trustworthiness, but only if the...
Model calibration, which is concerned with how frequently the model predicts correctly, not only pla...
Part 4: First Conformal Prediction and Its Applications Workshop (COPA 2012)International audienceVe...
Accurate uncertainty quantification is necessary to enhance the reliability of deep learning models ...
Deep neural networks are powerful tools to detect hidden patterns in data and leverage them to make ...
Successful use of probabilistic classification requires well-calibrated probability estimates, i.e.,...
Pretrained multilingual encoder models can directly perform zero-shot multilingual tasks or linguist...