
Improved Transformer Network Based on Multi scale Grouping Feature Fusion for Hyperspectral Image Classification
Author(s) -
Lin Wei,
Yuhang Liu,
Yuping Yin,
Yanning E,
Jinlong Gu
Publication year - 2025
Publication title -
ieee journal of selected topics in applied earth observations and remote sensing
Language(s) - English
Resource type - Magazines
SCImago Journal Rank - 1.246
H-Index - 88
eISSN - 2151-1535
pISSN - 1939-1404
DOI - 10.1109/jstars.2025.3596917
Subject(s) - geoscience , signal processing and analysis , power, energy and industry applications
In recent years, hyperspectral image (HSI) classification methods based on convolutional neural networks (CNNs) and Transformer architectures have achieved remarkable success. However, existing Transformer-based models exhibit several limitations, including insufficient capability in modeling local spectral structures, difficulty in capturing complex correlations between non-adjacent spectral bands, and excessive model parameter counts, which hinder their deployment in resource-constrained environments. To address these issues, this paper proposes a novel model, Multi-Scale Grouped Feature Fusion Vision Transformer (MGFF-ViT), designed to enhance spectral feature representation while maintaining parameter efficiency. First, we introduce a new spectral aggregation module, Spectral Odd-Even Fusion (SOEF), which reshapes spectral vectors into highly correlated spectral matrices. This allows the model to extract both local spectral features and long-range dependencies across frequency bands. Second, we propose a Multi-Scale Grouped Weight Attention (MGWA) mechanism that captures spatial contextual information at multiple scales. By applying a sorting-based strategy—“strong to strong, weak to weak”—the module strengthens relevant feature interactions while suppressing noise and redundancy. In addition to improving classification accuracy, MGFF-ViT significantly reduces the number of model parameters, making it lightweight and efficient for practical applications. Extensive experiments on four publicly available hyperspectral datasets demonstrate that MGFF-ViT achieves competitive or superior performance compared to state-of-the-art methods, even with limited training data. The code source will be available at https://github.com/liuyuhang456/MGFF-Net.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom