Recently, deformable convolutions based on convolutional neural networks have been widely used in hyperspectral image (HSI) classification due to their flexible geometric adaptability and superior local feature extraction capabilities. However, they still face significant challenges in establishing long-range dependencies and capturing global contextual information among pixel sequences. To address these challenges, a novel deformable convolution and Transformer hybrid network (DTHNet) is proposed for HSI classification. Specifically, PCA is firstly employed to reduce the dimensionality of the original HSI and a group depth joint convolution block (GDJCB) is utilized to capture the spectral-spatial features of the reduced HSI patches, avoiding the neglect of certain spectral bands. Secondly, a parallel architecture composed of a designed deformable convolution and a Transformer is utilized to jointly extract local-global spectral-spatial features and long-range dependencies in HSI. In the deformable convolution branch, a simple parameter-free attention (SimAM) enhanced spectral-spatial convolution block (SSCB) is designed to effectively prevent the loss of key information and the generation of redundant features during the convolution. In the Transformer branch, the deep integration of convolutional operation and self-attention mechanism further promotes more effective extraction of HSI features. Finally, fusion features from the two branches to obtain the more accurate HSI classification. Experimental results on three widely used HSI datasets demonstrate that the proposed DTHNet outperforms several state-of-the-art HSI classification networks.
扫码关注我们
求助内容:
应助结果提醒方式:
