Abstract

Satellite video is a novel data source for earth observation, which can be applied in multiple fields for dynamic monitoring. It is always equipped with high temporal resolution at the cost of low spatial resolution of tiny moving objects. Video super-resolution (VSR) is utilized to improve the spatial resolution of satellite video and obtain high spatial-temporal resolution data. However, most existing VSR methods mainly focus on the local inter-frame information during feature alignment, which lack the ability to model long-distance correspondence. In this article, a novel two-branch alignment network with an efficient fusion module is proposed for satellite VSR. Both deformable convolution and transformer-like attention are employed to fully explore the local and global information between frames. Further, a fusion module is proposed to model the residuals between fusion features and compensate them for better fusion. Experiments on Jilin-1 satellite videos demonstrate that the proposed network can achieve comparable results to current state-of-the-art VSR methods with tiny parameters.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call