Interpreting deep learning models in natural language processing: A review

X Sun, D Yang, X Li, T Zhang, Y Meng, H Qiu… - arXiv preprint arXiv …, 2021 - arxiv.org
Neural network models have achieved state-of-the-art performances in a wide range of
natural language processing (NLP) tasks. However, a long-standing criticism against neural …

DARE: disentanglement-augmented rationale extraction

L Yue, Q Liu, Y Du, Y An, L Wang… - Advances in Neural …, 2022 - proceedings.neurips.cc
Rationale extraction can be considered as a straightforward method of improving the model
explainability, where rationales are a subsequence of the original inputs, and can be …

FR: Folded rationalization with a unified encoder

W Liu, H Wang, J Wang, R Li, C Yue… - Advances in Neural …, 2022 - proceedings.neurips.cc
Rationalization aims to strengthen the interpretability of NLP models by extracting a subset
of human-intelligible pieces of their inputting texts. Conventional works generally employ a …

AGR: Reinforced causal agent-guided self-explaining rationalization

Y Zhao, Z Wang, X Li, J Liang, R Li - Proceedings of the 62nd …, 2024 - aclanthology.org
Most existing rationalization approaches are susceptible to degeneration accumulation due
to a lack of effective control over the learning direction of the model during training. To …

Decoupled rationalization with asymmetric learning rates: A flexible lipschitz restraint

W Liu, J Wang, H Wang, R Li, Y Qiu, Y Zhang… - Proceedings of the 29th …, 2023 - dl.acm.org
A self-explaining rationalization model is generally constructed by a cooperative game
where a generator selects the most human-intelligible pieces from the input text as …

Enhancing the rationale-input alignment for self-explaining rationalization

W Liu, H Wang, J Wang, Z Deng… - 2024 IEEE 40th …, 2024 - ieeexplore.ieee.org
Rationalization empowers deep learning models with self-explaining capabilities through a
cooperative game, where a generator selects a semantically consistent subset of the input …

D-separation for causal self-explanation

W Liu, J Wang, H Wang, R Li, Z Deng… - Advances in Neural …, 2024 - proceedings.neurips.cc
Rationalization aims to strengthen the interpretability of NLP models by extracting a subset
of human-intelligible pieces of their inputting texts. Conventional works generally employ the …

MGR: multi-generator based rationalization

W Liu, H Wang, J Wang, R Li, X Li, Y Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Rationalization is to employ a generator and a predictor to construct a self-explaining NLP
model in which the generator selects a subset of human-intelligible pieces of the input text to …

Mare: Multi-aspect rationale extractor on unsupervised rationale extraction

H Jiang, J Duan, Z Qu, J Wang - arXiv preprint arXiv:2410.03531, 2024 - arxiv.org
Unsupervised rationale extraction aims to extract text snippets to support model predictions
without explicit rationale annotation. Researchers have made many efforts to solve this task …

Learning Robust Rationales for Model Explainability: A Guidance-Based Approach

S Hu, K Yu - Proceedings of the AAAI Conference on Artificial …, 2024 - ojs.aaai.org
Selective rationalization can be regarded as a straightforward self-explaining approach for
enhancing model explainability in natural language processing tasks. It aims to provide …