TY - GEN
T1 - Interpreting undesirable pixels for image classification on black-box models
AU - Kang, Sin Han
AU - Jung, Hong Gyu
AU - Lee, Seong Whan
PY - 2019/10
Y1 - 2019/10
N2 - In an effort to interpret black-box models, researches for developing explanation methods have proceeded in recent years. Most studies have tried to identify input pixels that are crucial to the prediction of a classifier. While this approach is meaningful to analyse the characteristic of black-box models, it is also important to investigate pixels that interfere with the prediction. To tackle this issue, in this paper, we propose an explanation method that visualizes undesirable regions to classify an image as a target class. To be specific, we divide the concept of undesirable regions into two terms: (1) factors for a target class, which hinder that black-box models identify intrinsic characteristics of a target class and (2) factors for non-target classes that are important regions for an image to be classified as other classes. We visualize such undesirable regions on heatmaps to qualitatively validate the proposed method. Furthermore, we present an evaluation metric to provide quantitative results on ImageNet.
AB - In an effort to interpret black-box models, researches for developing explanation methods have proceeded in recent years. Most studies have tried to identify input pixels that are crucial to the prediction of a classifier. While this approach is meaningful to analyse the characteristic of black-box models, it is also important to investigate pixels that interfere with the prediction. To tackle this issue, in this paper, we propose an explanation method that visualizes undesirable regions to classify an image as a target class. To be specific, we divide the concept of undesirable regions into two terms: (1) factors for a target class, which hinder that black-box models identify intrinsic characteristics of a target class and (2) factors for non-target classes that are important regions for an image to be classified as other classes. We visualize such undesirable regions on heatmaps to qualitatively validate the proposed method. Furthermore, we present an evaluation metric to provide quantitative results on ImageNet.
KW - Explainable-AI
KW - Interpretability
UR - http://www.scopus.com/inward/record.url?scp=85082440554&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85082440554&partnerID=8YFLogxK
U2 - 10.1109/ICCVW.2019.00523
DO - 10.1109/ICCVW.2019.00523
M3 - Conference contribution
AN - SCOPUS:85082440554
T3 - Proceedings - 2019 International Conference on Computer Vision Workshop, ICCVW 2019
SP - 4250
EP - 4254
BT - Proceedings - 2019 International Conference on Computer Vision Workshop, ICCVW 2019
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 17th IEEE/CVF International Conference on Computer Vision Workshop, ICCVW 2019
Y2 - 27 October 2019 through 28 October 2019
ER -