Neural network architectures in natural language processing often use attention mechanisms to produce probability distributions over input token representations. Attention has empirically been demonstrated to improve performance in various tasks, while its weights have been extensively used as explanations for model predictions. Recent studies (Jain and Wallace, 2019; Serrano and Smith, 2019; Wiegreffe and Pinter, 2019) have showed that it cannot generally be considered as a faithful explanation (Jacovi and Goldberg, 2020) across encoders and tasks. In this paper, we seek to improve the faithfulness of attention-based explanations for text classification. We achieve this by proposing a new family of Task-Scaling (TaSc) mechanisms that learn...
This thesis focuses on model interpretability, an area concerned with under- standing model predicti...
Due to the black-box nature of deep learning models, methods for explaining the models’ results are ...
In the past decade, natural language processing (NLP) systems have come to be built almost exclusive...
International audienceAttention mechanism is contributing to the majority of recent advances in mach...
Attention is an increasingly popular mechanism used in a wide range of neural architectures. The mec...
Pretrained transformer-based models such as BERT have demonstrated state-of-the-art predictive perfo...
Explainability is one of the key factors in Natural Language Processing (NLP) specially for legal do...
Natural Language Inference (NLI) models are known to learn from biases and artefacts within their tr...
We propose an approach to faithfully explaining text classification models, using a specifically des...
Explainability is a key requirement for text classification in many application domains ranging from...
Attention mechanism has become a standard fixture in many state-of-the-art natural language processi...
Natural Language Inference (NLI) models are known to learn from biases and artefacts within their tr...
Deep neural networks (DNNs) can perform impressively in many natural language processing (NLP) tasks...
International audienceAttention maps in neural models for NLP are appealing to explain the decision ...
Neural network models with attention mechanism have shown their efficiencies on various tasks. Howev...
This thesis focuses on model interpretability, an area concerned with under- standing model predicti...
Due to the black-box nature of deep learning models, methods for explaining the models’ results are ...
In the past decade, natural language processing (NLP) systems have come to be built almost exclusive...
International audienceAttention mechanism is contributing to the majority of recent advances in mach...
Attention is an increasingly popular mechanism used in a wide range of neural architectures. The mec...
Pretrained transformer-based models such as BERT have demonstrated state-of-the-art predictive perfo...
Explainability is one of the key factors in Natural Language Processing (NLP) specially for legal do...
Natural Language Inference (NLI) models are known to learn from biases and artefacts within their tr...
We propose an approach to faithfully explaining text classification models, using a specifically des...
Explainability is a key requirement for text classification in many application domains ranging from...
Attention mechanism has become a standard fixture in many state-of-the-art natural language processi...
Natural Language Inference (NLI) models are known to learn from biases and artefacts within their tr...
Deep neural networks (DNNs) can perform impressively in many natural language processing (NLP) tasks...
International audienceAttention maps in neural models for NLP are appealing to explain the decision ...
Neural network models with attention mechanism have shown their efficiencies on various tasks. Howev...
This thesis focuses on model interpretability, an area concerned with under- standing model predicti...
Due to the black-box nature of deep learning models, methods for explaining the models’ results are ...
In the past decade, natural language processing (NLP) systems have come to be built almost exclusive...