Abstract
An exponential growth of interest in the healthcare IoT over the past few years has increased the adoption of AI. However, healthcare analytics demands highly accurate and reliable systems that can not afford even a small amount of vulnerability. The black-box nature of AI models raises significant concerns about the computations involved in these predictions, leading to a lack of trust in these models. Explainable Artificial Intelligence (XAI) seeks to resolve this problem by providing a human-understandable explanation for AI decisions, bringing transparency, trust, and fairness to these AI models. This paper evaluates the existing XAI models on numeric and graph-structured healthcare data. The LIME and SHAP models are first used to explain the predictions in the numerical dataset of fetal health prediction. Second, graph-structured data is analyzed using GNN Explainer and PG Explainer. This evaluation suggests that XAI model interpretations benefit healthcare professionals and patients to trust AI predictions.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.