Abstract
Most vision-language tasks, such as image captioning, require complex deep learning (DL) models/architectures. However, the high complexity of such models often leads their inner functioning to be regarded as black-boxes, preventing humans from understanding their decision process. Representation space is a key element in DL models, understanding this space and how information is encoded constitute an interesting approach for interpretable DL. Indeed, as far as we are aware, this paper is the first one presenting a novel method based on perturbation principle of representation space, to study the components that influence DL architectures used in image captioning. The core idea is to isolate and identify the importance of each component (element of the architecture) involved in the captioning pipeline by perturbing, by means of Gaussian functions, the representation space rather than the original space of inputs. We experimentally demonstrate that those components differ in their influence and relevance. The results show that the visual modality would constitute a critical explanation target in captioning models in contrast to language modality, thus leading to more fine-grained explanations. We also propose MSICE, an automatic evaluation metric for image captioning which addresses two important yet overlooked linguistic aspects, morphology and semantic. Our code will be publicly accessible to support future research in this field.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.