作者
Srihari Maruthi, Sarath Babu Dodda, Ramswaroop Reddy Yellu, Praveen Thuniki, Surendranadha Reddy Byrapu Reddy
发表日期
2022/12/31
来源
Australian Journal of Machine Learning Research & Applications
卷号
2
期号
2
页码范围
1-9
简介
Language models have achieved remarkable success in various natural language processing tasks, but their complex inner workings often lack transparency, leading to concerns about their reliability and ethical implications. Explainable AI (XAI) methods aim to address this issue by providing insights into how language models make decisions. This paper presents a comprehensive review of XAI methods for interpreting and explaining the decisions made by language models. We discuss key approaches such as attention mechanisms, saliency maps, and model-agnostic techniques, highlighting their strengths and limitations. Additionally, we explore the implications of XAI for enhancing the transparency and trustworthiness of language models in real-world applications.
引用总数