Explainable AI (XAI) promotes understandability and credibility in the complex decision-making process of machine learning models. Post-hoc explanations for deep neural networks are crucial because they focus on faithful explanation of the learned representations, decision mechanisms, and uncertainty of the model. Explaining deep convolutional neural networks (DCNN) is particularly challenging because of the high dimensionality of deep features and the complexity of model inference. Most post-hoc explanation methods follow one single paradigm of explanation, confining the diversity and consistency of the explanation. This study proposes a joint Gaussian mixture model (JGMM), which is a probabilistic model that jointly models deep inter-layer features and produces faithful and consistent post-hoc explanations. The JGMM explains deep features by Gaussian mixture model and model inference by the posterior distribution of the latent component variables. JGMM enables a versatile explanation framework that unifies the interpretable proxy model and mining global and local explanatory examples. Experiments are performed on various DCNN image classifiers in comparison with other explaining methods. The results show that the JGMM can efficiently produce versatile, consistent, faithful, and understandable explanations.
Read full abstract