Fault-Line Image Explanations


We propose this new explanation framework to identify the minimal semantic-level features (e.g. stripes on zebra, pointed ears on dog) referred to as the explainable concept, that can be added/deleted to alter the classification category.

Intended Use

Our framework is helpful for generating conceptual and counterfactual explanations for an image classification model.

The proposed framework is applicable to image analytics domain.


We take a set of input images and extract xconcepts. These xconcepts are then utilized in generating optimal explanations.


  1. https://ojs.aaai.org/index.php/AAAI/article/view/5643
  2. https://vcla.stat.ucla.edu/Projects/ArjunXAI/CoCoX.html
  3. https://github.com/arjunakula/arjunakula.github.io/blob/master/AAAI2020_poster%20(1).pdf
  4. https://www.dropbox.com/s/kde88lgv78v8dyj/Arjun_UCLA_Analytics_v2.pptx?dl=0