Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/27115
Title: Ultralightweight Spatial-Spectral Feature Cooperation Network for Change Detection in Remote Sensing Images
Authors: Lei, T
Geng, X
Ning, H
Lv, Z
Gong, M
Jin, Y
Nandi, AK
Keywords: change detection (CD);convolutional neural network (CNN);multiscale feature extraction;spatial–spectral feature cooperation (SSFC)
Issue Date: 24-Mar-2023
Publisher: Institute of Electrical and Electronics Engineers (IEEE)
Citation: Lei, T. et al. (2023) 'Ultralightweight Spatial-Spectral Feature Cooperation Network for Change Detection in Remote Sensing Images', IEEE Transactions on Geoscience and Remote Sensing, 61, 4402114, pp. 1 - 14. doi: 10.1109/TGRS.2023.3261273.
Abstract: Copyright © The Author(s) 2023. Deep convolutional neural networks (CNNs) have achieved much success in remote sensing image change detection (CD) but still suffer from two main problems. First, the existing multiscale feature fusion methods often use redundant feature extraction and fusion strategies, which often lead to high computational costs and memory usage. Second, the regular attention mechanism in CD is difficult to model spatial–spectral features and generate 3-D attention weights at the same time, ignoring the cooperation between spatial features and spectral features. To address the above issues, an efficient ultralightweight spatial–spectral feature cooperation network (USSFC-Net) is proposed for CD in this article. The proposed USSFC-Net has two main advantages. First, a multiscale decoupled convolution (MSDConv) is designed, which is clearly different from the popular atrous spatial pyramid pooling (ASPP) module and its variants since it can flexibly capture the multiscale features of changed objects using cyclic multiscale convolution. Meanwhile, the design of MSDConv can greatly reduce the number of parameters and computational redundancy. Second, an efficient spatial–spectral feature cooperation (SSFC) strategy is introduced to obtain richer features. The SSFC differs from the existing 2-D attention mechanisms since it learns 3-D spatial–spectral attention weights without adding any parameters. The experiments on three datasets for remote sensing image CD demonstrate that the proposed USSFC-Net achieves better CD accuracy than most CNNs-based methods and requires lower computational costs and fewer parameters, even it is superior to some Transformer-based methods. The code is available at https://github.com/SUST-reynole/USSFC-Net .
URI: https://bura.brunel.ac.uk/handle/2438/27115
DOI: https://doi.org/10.1109/TGRS.2023.3261273
ISSN: 0196-2892
Other Identifiers: ORCID iDs: Tao Lei https://orcid.org/0000-0002-2104-9298; Xinzhe Geng https://orcid.org/0009-0007-0950-3957; Hailong Ning https://orcid.org/0000-0001-8375-1181; Zhiyong Lv https://orcid.org/0000-0003-2595-4794; Maoguo Gong https://orcid.org/0000-0002-0415-8556; Yaochu Jin https://orcid.org/0000-0003-1100-0631; Asoke K. Nandi https://orcid.org/0000-0001-6248-2875.
4402114
Appears in Collections:Dept of Electronic and Electrical Engineering Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © The Author(s) 2023. Published by Institute of Electrical and Electronics Engineers (IEEE). This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4.0/2.95 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons