Local Interpretations for Explainable Natural Language Processing: A Survey
ACM Computing Surveys(2021)
摘要
As the use of deep learning techniques has grown across various fields over
the past decade, complaints about the opaqueness of the black-box models have
increased, resulting in an increased focus on transparency in deep learning
models. This work investigates various methods to improve the interpretability
of deep neural networks for Natural Language Processing (NLP) tasks, including
machine translation and sentiment analysis. We provide a comprehensive
discussion on the definition of the term interpretability and its various
aspects at the beginning of this work. The methods collected and summarised in
this survey are only associated with local interpretation and are specifically
divided into three categories: 1) interpreting the model's predictions through
related input features; 2) interpreting through natural language explanation;
3) probing the hidden states of models and word representations.
更多查看译文
关键词
Deep neural networks,explainable AI,local interpretation,natural language processing
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要