Spatial and polarimetric information fusion using residual network for polarimetric synthetic aperture radar image classification
Author(s) -
Imani Maryam
Publication year - 2022
Publication title -
iet radar, sonar and navigation
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.489
H-Index - 82
eISSN - 1751-8792
pISSN - 1751-8784
DOI - 10.1049/rsn2.12310
Subject(s) - polarimetry , synthetic aperture radar , residual , artificial intelligence , computer science , remote sensing , radar imaging , computer vision , pattern recognition (psychology) , radar , geology , algorithm , telecommunications , physics , optics , scattering
A polarimetric synthetic aperture radar (POLSAR) system provides an image that can be considered as a data cube containing spatial information in two spatial dimensions and polarimetric information in the scattering dimension. A spatial and polarimetric residual network (SPRN) is proposed for POLSAR image classification. At first, polarimetric features are extracted from the scattering dimension through two designed polarimetric residual blocks. Then, the processed POLSAR cube is fed to two consecutive spatial residual blocks for contextual feature extraction. Three dimensional convolutional layers are used as basic layers for simultaneous extraction and fusion of polarimetric information and correlation among neighbouring pixels in local regions. The shortcut connections are utilised to overcome the degradation problem due to increasing network depth. In addition, batch normalisation is applied to regularise the learning process. The experimental results on four real POLSAR images show the superior performance of SPRN compared to several state‐of‐the‐art classifiers in terms of various assessment measures.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom