Chain Graph Explanation of Neural Network Based on Feature-Level Class Confusion

Hyekyoung Hwang, Eunbyung Park, Jitae Shin

Research output: Contribution to journalArticlepeer-review

3 Scopus citations

Abstract

Despite increasing interest in developing interpretable machine learning methods, most recent studies have provided explanations only for single instances, require additional datasets, and are sensitive to hyperparameters. This paper proposes a confusion graph that reveals model weaknesses by constructing a confusion dictionary. Unlike other methods, which focus on the performance variation caused by single-neuron suppression, it defines the role of each neuron in two different perspectives: ‘correction’ and ‘violation.’ Furthermore, our method can identify the class relationships in similar positions at the feature level, which can suggest improvements to the model. Finally, the proposed graph construction is model-agnostic and does not require additional data or tedious hyperparameter tuning. Experimental results show that the information loss from omitting the channels guided by the proposed graph can result in huge performance degradation, from 91% to 33%, while the proposed graph only retains 1% of total neurons.

Original languageEnglish
Article number1523
JournalApplied Sciences (Switzerland)
Volume12
Issue number3
DOIs
StatePublished - 1 Feb 2022
Externally publishedYes

Keywords

  • Computer vision
  • Convolution neural network
  • Deep learning
  • Explainable AI

Fingerprint

Dive into the research topics of 'Chain Graph Explanation of Neural Network Based on Feature-Level Class Confusion'. Together they form a unique fingerprint.

Cite this