Working Paper - Comparing explainability through similarity

Image credit: F.Gwinner

Abstract

In our work, we propose the use of Representational Similarity Analysis (RSA) for explainable AI (XAI) approaches, intending to gain further insights into Machine Learning (ML) models as well as their explanations. We created a computational evaluative study to demonstrate that Similarity Analysis of explanations can contribute to a better understanding of XAI-based decision support systems. In our study, we address how the application of our approach can be leveraged to analyze the robustness of explanations for several different underlying types of changes in the ML Pipeline. The study results further indicate that changes in the ML pipeline, such as an altered preprocessing or a different machine learning model, lead to different explanations and to which extent the similarity changes. Similarity analysis of explanations allows practitioners to monitor explanation robustness and compare different XAI systems for the same decisions.Conjointly to the results and practical use in operationalized ML, as well as the bene- fits and limitations of the approach, we discuss insights into the field of computational neuroscience and neural information processing

Publication
Working Paper - Comparing explainability through similarity
Fabian Gwinner
Fabian Gwinner
PHD Candidate

My research interests include Machine Learning, especially Deep Learning, XAI, Decision Support Systems