Authors
Pieter-Jan Kindermans, Sara Hooker, Julius Adebayo, Maximilian Alber, Kristof T Schütt, Sven Dähne, Dumitru Erhan, Been Kim
Publication date
2019
Journal
Explainable AI: Interpreting, explaining and visualizing deep learning
Pages
267-280
Publisher
Springer International Publishing
Description
Saliency methods aim to explain the predictions of deep neural networks. These methods lack reliability when the explanation is sensitive to factors that do not contribute to the model prediction. We use a simple and common pre-processing step which can be compensated for easily—adding a constant shift to the input data—to show that a transformation with no effect on how the model makes the decision can cause numerous methods to attribute incorrectly. In order to guarantee reliability, we believe that the explanation should not change when we can guarantee that two networks process the images in identical manners. We show, through several examples, that saliency methods that do not satisfy this requirement result in misleading attribution. The approach can be seen as a type of unit test; we construct a narrow ground truth to measure one stated desirable property. As such, we hope the community will …
Total citations
201720182019202020212022202320242395898120168166104
Scholar articles
PJ Kindermans, S Hooker, J Adebayo, M Alber… - Explainable AI: Interpreting, explaining and visualizing …, 2019
PJ Kindermans, S Hooker, J Adebayo, M Alber… - arXiv preprint arXiv:1711.00867, 2018
PJ Kindermans, S Hooker, J Adebayo, M Alber… - arXiv preprint arXiv:1711.00867
PJ Kindermans, S Hooker, J Adebayo, M Alber… - arXiv preprint arXiv:1711.00867, 2017