The use of counterfactual explanations (CFXs) is an increasingly popular explanation strategy for machine learning models. However, recent studies have shown that these explanations may not be robust to changes in the underlying model (e.g., following retraining), which raises questions about their reliability in real-world applications. Existing attempts towards solving this problem are heuristic, and the robustness to model changes of the resulting CFXs is evaluated with only a small number of retrained models, failing to provide exhaustive guarantees. To remedy this, we propose the first notion to formally and deterministically assess the robustness (to model changes) of CFXs for neural networks, that we call ∆-robustness. We introduce a...
We present VeriX, a system for producing optimal robust explanations and generating counterfactuals ...
As artificial intelligence (AI) systems increasingly impact the society, it is important to design a...
The explainable AI literature contains multiple notions of what an explanation is and what desiderat...
The use of counterfactual explanations (CFXs) is an increasingly popular explanation strategy for ma...
Counterfactual Explanations (CEs) have received increasing interest as a major methodology for expla...
Counterfactual explanations are a prominent example of post-hoc interpretability methods in the expl...
Counterfactual explanations inform ways to achieve a desired outcome from a machine learning model. ...
We propose a novel method for explaining the predictions of any classifier. In our approach, local e...
Counterfactual explanations (CEs) are a powerful means for understanding how decisions made by algor...
Robustness and counterfactual bias are usually evaluated on a test dataset. However, are these evalu...
In contrast with ad-hoc methods for eXplainable Artificial Intelligence (XAI), formal explainability...
Massive deployment of Graph Neural Networks (GNNs) in high-stake applications generates a strong dem...
International audienceCounterfactual explanations have become a mainstay of the XAI field. This part...
Counterfactual explanations (CFEs) exemplify how to minimally modify a feature vector to achieve a d...
Counterfactual explanations (CEs) are a powerful means for understanding how decisions made by algor...
We present VeriX, a system for producing optimal robust explanations and generating counterfactuals ...
As artificial intelligence (AI) systems increasingly impact the society, it is important to design a...
The explainable AI literature contains multiple notions of what an explanation is and what desiderat...
The use of counterfactual explanations (CFXs) is an increasingly popular explanation strategy for ma...
Counterfactual Explanations (CEs) have received increasing interest as a major methodology for expla...
Counterfactual explanations are a prominent example of post-hoc interpretability methods in the expl...
Counterfactual explanations inform ways to achieve a desired outcome from a machine learning model. ...
We propose a novel method for explaining the predictions of any classifier. In our approach, local e...
Counterfactual explanations (CEs) are a powerful means for understanding how decisions made by algor...
Robustness and counterfactual bias are usually evaluated on a test dataset. However, are these evalu...
In contrast with ad-hoc methods for eXplainable Artificial Intelligence (XAI), formal explainability...
Massive deployment of Graph Neural Networks (GNNs) in high-stake applications generates a strong dem...
International audienceCounterfactual explanations have become a mainstay of the XAI field. This part...
Counterfactual explanations (CFEs) exemplify how to minimally modify a feature vector to achieve a d...
Counterfactual explanations (CEs) are a powerful means for understanding how decisions made by algor...
We present VeriX, a system for producing optimal robust explanations and generating counterfactuals ...
As artificial intelligence (AI) systems increasingly impact the society, it is important to design a...
The explainable AI literature contains multiple notions of what an explanation is and what desiderat...