BriefGPT.xyz
Dec, 2016
通过表示擦除理解神经网络
Understanding Neural Networks through Representation Erasure
HTML
PDF
Jiwei Li, Will Monroe, Dan Jurafsky
TL;DR
本文提出了一种使用擦除来解释神经模型决策的通用方法,通过分析和比较擦除输入词向量的不同部分,中间隐藏层的不同单元或输入词的不同方法,来评估和解释模型判断的影响。研究表明,该方法不仅可以提供清晰的神经模型决策解释,还可以为神经模型的错误分析提供一种方法。
Abstract
While
neural networks
have been successfully applied to many natural language processing tasks, they come at the cost of
interpretability
. In this paper, we propose a general methodology to analyze and interpret
→