Abstract

Tradeoffs among the spatial, spectral, and temporal resolutions of satellite sensors make it difficult to acquire remote sensing images at both high spatial and high temporal resolutions from an individual sensor. Studies have developed methods to fuse spatiotemporal data from different satellite sensors, and these methods often assume linear changes in surface reflectance across time and adopt empirical rules and handcrafted features. Here, we propose a dense spatiotemporal fusion (DenseSTF) network based on the convolutional neural network (CNN) to deal with these problems. DenseSTF uses a patch-to-pixel modeling strategy that can provide abundant texture details for each pixel in the target fine image to handle heterogeneous landscapes and models both forward and backward temporal dependencies to account for land cover changes. Moreover, DenseSTF adopts a mapping function with few assumptions and empirical rules, which allows for establishing reliable relationships between the coarse and fine images. We tested DenseSTF in three contrast scenes with different degrees of heterogeneity and temporal changes, and made comparisons with three rule-based fusion approaches and three CNNs. Experimental results indicate that DenseSTF can provide accurate fusion results and outperform the other tested methods, especially when the land cover changes abruptly. The structure of the deep learning networks largely impacts the success of data fusion. Our study developed a novel approach based on CNN using a patch-to-pixel mapping strategy and highlighted the effectiveness of the deep learning networks in the spatiotemporal fusion of the remote sensing data.

Highlights

  • THE fast development of modern satellite technology has advanced the usages of long-term time series of remote sensing images in the monitoring and modeling of the land surface processes [1,2,3]

  • Visual inspection suggests that the fusion image produced by DenseSTF is highly consistent with the observed Landsat image

  • Compared with spatiotemporal fusion network (STFNET) and enhanced DCSTFN (EDCSTFN), the fused image by VGG16 is closer to the observed Landsat image

Read more

Summary

Introduction

THE fast development of modern satellite technology has advanced the usages of long-term time series of remote sensing images in the monitoring and modeling of the land surface processes [1,2,3]. Given the differences in satellite sensors, orbit altitudes, and revisiting periods, there are tradeoffs regarding the spatial, temporal, and spectral resolutions of images acquired from an individual satellite sensor [4]. Moderate Resolution Imaging Spectroradiometer (MODIS) provides observations at the spatial resolution ranging from 250 to 1000 m and has a revisit time of nearly one day for most areas across the world. Landsat acquires images at high spatial resolution of 30 m but relatively small scene coverage, and its revisit time is up to 16 days. There are large application demands that require continuous remote sensing data at both high spatial and temporal resolutions. Developing spatiotemporal fusion methods to blend remote sensing data acquired by different satellite sensors has become a research frontier in the field of remote sensing [7,8,9]

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call