Semantic Scholar Open Access 2019 3025 sitasi

Interpretable Machine Learning

Bradley C. Boehmke Brandon M. Greenwell

Abstrak

Interpretable machine learning has become a popular research direction as deep neural networks (DNNs) have become more powerful and their applications more mainstream, yet DNNs remain difficult to understand. Testing with Concept Activation Vectors, TCAV, (Kim et al. 2017) is an approach to interpreting DNNs in a human-friendly way and has recently received significant attention in the machine learning community. The TCAV algorithm achieves a degree of global interpretability for DNNs through human-defined concepts as explanations. This project introduces Robust TCAV, which builds on TCAV and experimentally determines best practices for this method. The objectives for Robust TCAV are 1) Making TCAV more consistent by reducing variance in the TCAV score distribution and 2) Increasing CAV and TCAV score resistance to perturbations. A difference of means method for CAV generation was determined to be the best practice to achieve both objectives. Many areas of the TCAV process are explored including CAV visualization in low dimensions, negative class selection, and activation perturbation in the direction of a CAV. Finally, a thresholding technique is considered to remove noise in TCAV scores. This project is a step in the direction of making TCAV, an already impactful algorithm in interpretability, more reliable and useful for practitioners.

Topik & Kata Kunci

Penulis (2)

B

Bradley C. Boehmke

B

Brandon M. Greenwell

Format Sitasi

Boehmke, B.C., Greenwell, B.M. (2019). Interpretable Machine Learning. https://doi.org/10.1201/9780367816377-16

Akses Cepat

Lihat di Sumber doi.org/10.1201/9780367816377-16
Informasi Jurnal
Tahun Terbit
2019
Bahasa
en
Total Sitasi
3025×
Sumber Database
Semantic Scholar
DOI
10.1201/9780367816377-16
Akses
Open Access ✓