Abstract

The Visible Infrared Imaging Radiometer Suite (VIIRS) onboard the Suomi National Polar-orbiting Partnership (S-NPP) satellite has been used for the early detection and daily monitoring of active wildfires. How to effectively segment the active fire pixels from VIIRS image time-series in a reliable manner remains a challenge because of the low precision associated with high recall using automatic methods. For active fire detection, multi-criteria thresholding is often applied to both low-resolution and mid-resolution Earth observation images. Deep learning approaches based on Convolutional Neural Networks are also well-studied on mid-resolution images. However, ConvNet-based approaches have poor performance on low-resolution images because of the coarse spatial features. On the other hand, the high temporal resolution of VIIRS images highlights the potential of using sequential models for active fire detection. Transformer networks, a recent deep learning architecture based on self-attention, offer hope as they have shown strong performance on image segmentation and sequential modelling tasks within computer vision. In this research, we propose a Transformer-based solution to segment active fire pixels from the VIIRS time-series. The solution feeds a time-series of tokenized pixels into a Transformer network to identify active fire pixels at each timestamp and achieves a significantly higher F1-Score than prior approaches for active fires within the study areas in California, New Mexico, and Oregon in the US, and in British Columbia and Alberta in Canada, as well as in Australia and Sweden.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call