Abstract
Medical images are difficult to comprehend for a person without expertise. The scarcity of medical practitioners across the globe often face the issue of physical and mental fatigue due to the high number of cases, inducing human errors during the diagnosis. In such scenarios, having an additional opinion can be helpful in boosting the confidence of the decision maker. Thus, it becomes crucial to have a reliable visual question answering (VQA) system to provide a ‘second opinion’ on medical cases. However, most of the VQA systems that work today cater to real-world problems and are not specifically tailored for handling medical images. Moreover, the VQA system for medical images needs to consider a limited amount of training data available in this domain. In this paper, we develop MedFuseNet, an attention-based multimodal deep learning model, for VQA on medical images taking the associated challenges into account. Our MedFuseNet aims at maximizing the learning with minimal complexity by breaking the problem statement into simpler tasks and predicting the answer. We tackle two types of answer prediction—categorization and generation. We conducted an extensive set of quantitative and qualitative analyses to evaluate the performance of MedFuseNet. Our experiments demonstrate that MedFuseNet outperforms the state-of-the-art VQA methods, and that visualization of the captured attentions showcases the intepretability of our model’s predicted results.
Highlights
Medical images are difficult to comprehend for a person without expertise
We quantitatively evaluate the performance of MedFuseNet and compare it with the baseline models described in the “visual question answering (VQA) baseline models for comparison” section for the tasks of answer categorization and answer generation
Whereas the Bilinear Attention Networks (BAN) model is more competitive to MedFuseNet model for category 3, while the BAN model under-performs our model for category 1 by 2 percent and category 2 by 1.4 percent
Summary
We first provide an overview of related works for VQA tasks for real-world and medical domains, and discuss the related works on components of VQA approaches. A typical model for VQA first extracts the feature vectors from multiple modalities (image and question text), and combines the vectors using any one of the above-stated fusion techniques, and predicts the answer from the fused vector. It uses this attended vector as an input to the image attention mechanism as described in Algorithm 1 from lines 8-18, instead of question feature vector q. As shown in Algorithm 1 (lines 1–12), the MedFuseNet first extracts the feature vectors vand qfor input image v and question q, respectively This is followed by the computation of the attended question features qe using question attention mechanism Eq(q).
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.