Abstract
Graph Neural Networks (GNN) are powerful tools for deep learning. Similar to other neural networks, GNNs are complex models, in which humans can’t understand the decision-making procedures of the models. Therefore, it brings the need to explainability of GNNs. Explainability is critical for deep learning to support its predictions. In this paper, we will investigate the Grad-Cam and Integrated-Gradients explaining methods. The Grad-Cam applies a global average pooling over the feature activation mapping, and then which was followed by a ReLU activation to obtain an attribution. The Integrated-Gradients explains models by taking a line integral between the baseline image (a black image) and the source image. We demonstrate how Grad-Cam and the Integrated-Gradients methods explain the graph-deep model in semantic segmentation tasks over the Cityscapes dataset. FCN and LRASSP-MobileNet are used as a comparison to the DualGCN in the experiment to show the explaining effect.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.