Abstract

In recent years, convolutional neural networks (CNNs) have drawn significant attention for classification of hyperspectral images (HSIs). Due to their self-attention mechanism, the vision transformer (ViT) provides promising classification performance compared to CNNs. Many researchers have incorported ViT for HSI classification purposes. However, its performance can be further improved because the current version does not use spatial-spectral features. In this paper, we present a new morphological transformer (morphFormer) that implements a learnable spectral and spatial morphological network, where spectral and spatial morphological convolution operations are used (in conjunction with the attention mechanism) to improve the interaction between the structural and shape information of the HSI token and the CLS token. Experiments conducted on widely used HSIs demonstrate the superiority of proposed morphFormer over the classical CNN models and state-of-the-art transformer models. The source will be made available publicly at https://github.com/mhaut/morphFormer.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call