Abstract
Much of the recent work in remote sensing image captioning is influenced by natural image captioning. These methods tend to fix the defects of the model architecture to improve the previous work, but pay little attention to the differences between remote sensing images and natural images. By considering these differences, we propose a multiscale multiinteraction remote sensing image captioning model. As in Fig. 1(a), the targets in remote sensing images have a wide range of scales; while the natural images are generally taken close-up, resulting in a similar scale for the foreground targets. Due to the difference in shooting methods, the model pretrained on close-up natural images cannot capture multiscale remote sensing targets well. To alleviate this problem, we propose a two-stage multiscale structure for feature representation, where we first finetune the CNN backbone on remote sensing images for domain adaption, then we collect features from different stages as the multiscale feature representation. Moreover, due to the shooting distance, the height information of the target in the remote sensing image is greatly weakened, thus some objects like low plants and grasses become difficult to identify, as in Fig. 1(b). Thus, we further propose a multiinteraction feature representation module, where information flow of the same and different layers could effectively interact. By calculating the similarity score among features, we fuse features with high similarity, and increase the distance between features of different categories, thereby enhancing the distinguishability. Results on RSICD, Sydney-Captions, and UCM-Captions show a clear improvement over the compared methods.
Highlights
I MAGE captioning lies in the intersection of natural language processing (NLP) and computer vision (CV), which is used to generate natural language description for the input image based on its content
Lu et al [2] collected a larger dataset for remote sensing image captioning, which is more in line with the characteristics of remote sensing data and they tested a variety of baseline models borrowed from natural image captioning task
As previous work tends to adopt similar methods as natural images to perform remote sensing image captioning (e.g., VGG [14], ResNet [15] pre-trained on natural images are used for image features extraction, LSTM [16] is further applied for text generation), a question arises: Is remote sensing image captioning the same as natural image captioning? We argue that some characteristics of remote sensing images should be considered when we describe them: 1) Backbone networks trained on natural images are generally for close-up shooting of the target objects, so the network mainly focuses on objects that occupy most space of the image
Summary
I MAGE captioning lies in the intersection of natural language processing (NLP) and computer vision (CV), which is used to generate natural language description for the input image based on its content. Different from the low-level vision tasks, such as remote sensing image classification [4], [5], object/saliency detection [6], [7] and semantic/instance segmentation [8], [9], remote sensing image captioning requires identifying the categories and regions of objects in the image, and inferring relation of them, and fluently expressed in natural language. Inspired by natural image captioning methods [10]–[12], many recent works [1], [2], [13] adopted the basic encoder-decoder pipeline for remote sensing image captioning, e.g., Qu et al [1] first proposed to leverage the CNN-RNN architecture to describe the remote sensing images. Lu et al [2] collected a larger dataset for remote sensing image captioning, which is more in line with the characteristics of remote sensing data and they tested a variety of baseline models borrowed from natural image captioning task
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.