arXiv Open Access 2019

Interpreting Undesirable Pixels for Image Classification on Black-Box Models

Sin-Han Kang Hong-Gyu Jung Seong-Whan Lee
Lihat Sumber

Abstrak

In an effort to interpret black-box models, researches for developing explanation methods have proceeded in recent years. Most studies have tried to identify input pixels that are crucial to the prediction of a classifier. While this approach is meaningful to analyse the characteristic of blackbox models, it is also important to investigate pixels that interfere with the prediction. To tackle this issue, in this paper, we propose an explanation method that visualizes undesirable regions to classify an image as a target class. To be specific, we divide the concept of undesirable regions into two terms: (1) factors for a target class, which hinder that black-box models identify intrinsic characteristics of a target class and (2) factors for non-target classes that are important regions for an image to be classified as other classes. We visualize such undesirable regions on heatmaps to qualitatively validate the proposed method. Furthermore, we present an evaluation metric to provide quantitative results on ImageNet.

Topik & Kata Kunci

Penulis (3)

S

Sin-Han Kang

H

Hong-Gyu Jung

S

Seong-Whan Lee

Format Sitasi

Kang, S., Jung, H., Lee, S. (2019). Interpreting Undesirable Pixels for Image Classification on Black-Box Models. https://arxiv.org/abs/1909.12446

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2019
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓