Abstract

Local Interpretable Model-Agnostic Explanations (LIME) is a popular technique used to increase the interpretability and explainability of black box Machine Learning (ML) algorithms. LIME typically creates an explanation for a single prediction by any ML model by learning a simpler interpretable model (e.g., linear classifier) around the prediction through generating simulated data around the instance by random perturbation, and obtaining feature importance through applying some form of feature selection. While LIME and similar local algorithms have gained popularity due to their simplicity, the random perturbation methods result in shifts in data and instability in the generated explanations, where for the same prediction, different explanations can be generated. These are critical issues that can prevent deployment of LIME in sensitive domains. We propose a deterministic version of LIME. Instead of random perturbation, we utilize Agglomerative Hierarchical Clustering (AHC) to group the training data together and K-Nearest Neighbour (KNN) to select the relevant cluster of the new instance that is being explained. After finding the relevant cluster, a simple model (i.e., linear model or decision tree) is trained over the selected cluster to generate the explanations. Experimental results on six public (three binary and three multi-class) and six synthetic datasets show the superiority for Deterministic Local Interpretable Model-Agnostic Explanations (DLIME), where we quantitatively determine the stability and faithfulness of DLIME compared to LIME.

Highlights

  • Published: 30 June 2021Recent decades have witnessed the rise of Artificial Intelligence (AI) and MachineLearning (ML) in critical domains such as healthcare, criminal justice and finance [1]

  • We propose an array of Deterministic Local Interpretable Model-Agnostic Explanations (DLIME) frameworks

  • DLIME uses Agglomerative Hierarchical Clustering (AHC) by utilizing K-Nearest Neighbour (KNN) to find cluster of data points that are similar to a test instance

Read more

Summary

Introduction

Published: 30 June 2021Recent decades have witnessed the rise of Artificial Intelligence (AI) and MachineLearning (ML) in critical domains such as healthcare, criminal justice and finance [1]. Sometimes the binary “yes” or “no” answer is not sufficient and questions such as “how” or “where” something occurred is more significant. A few interpretable and explainable models have been proposed in recent literature. These approaches can be grouped based on different criterion [1,2,3,4] such as: (i) Model agnostic or model specific; (ii) Local or global; (iii) Local instance-wise or group-wise; (iv) Intrinsic or post hoc; (v) Variable importance or sensitivity analysis; (vi) Features importance or saliency mapping. LIME is an instance-based explainer, which generates simulated data points around an instance through random perturbation, and provides explanations by fitting a weighted sparse linear model over predicted responses from the perturbed points. The random perturbation may Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations

Objectives
Methods
Results
Discussion
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.