z-logo
open-access-imgOpen Access
Improved Transformer Network Based on Multi scale Grouping Feature Fusion for Hyperspectral Image Classification
Author(s) -
Lin Wei,
Yuhang Liu,
Yuping Yin,
Yanning E,
Jinlong Gu
Publication year - 2025
Publication title -
ieee journal of selected topics in applied earth observations and remote sensing
Language(s) - English
Resource type - Magazines
SCImago Journal Rank - 1.246
H-Index - 88
eISSN - 2151-1535
pISSN - 1939-1404
DOI - 10.1109/jstars.2025.3596917
Subject(s) - geoscience , signal processing and analysis , power, energy and industry applications
In recent years, hyperspectral image (HSI) classification methods based on convolutional neural networks (CNNs) and Transformer architectures have achieved remarkable success. However, existing Transformer-based models exhibit several limitations, including insufficient capability in modeling local spectral structures, difficulty in capturing complex correlations between non-adjacent spectral bands, and excessive model parameter counts, which hinder their deployment in resource-constrained environments. To address these issues, this paper proposes a novel model, Multi-Scale Grouped Feature Fusion Vision Transformer (MGFF-ViT), designed to enhance spectral feature representation while maintaining parameter efficiency. First, we introduce a new spectral aggregation module, Spectral Odd-Even Fusion (SOEF), which reshapes spectral vectors into highly correlated spectral matrices. This allows the model to extract both local spectral features and long-range dependencies across frequency bands. Second, we propose a Multi-Scale Grouped Weight Attention (MGWA) mechanism that captures spatial contextual information at multiple scales. By applying a sorting-based strategy—“strong to strong, weak to weak”—the module strengthens relevant feature interactions while suppressing noise and redundancy. In addition to improving classification accuracy, MGFF-ViT significantly reduces the number of model parameters, making it lightweight and efficient for practical applications. Extensive experiments on four publicly available hyperspectral datasets demonstrate that MGFF-ViT achieves competitive or superior performance compared to state-of-the-art methods, even with limited training data. The code source will be available at https://github.com/liuyuhang456/MGFF-Net.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom