1
|
Sun Y, Lei L, Guan D, Kuang G, Li Z, Liu L. Locality Preservation for Unsupervised Multimodal Change Detection in Remote Sensing Imagery. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:6955-6969. [PMID: 38809739 DOI: 10.1109/tnnls.2024.3401696] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2024]
Abstract
Multimodal change detection (MCD) is a topic of increasing interest in remote sensing. Due to different imaging mechanisms, the multimodal images cannot be directly compared to detect the changes. In this article, we explore the topological structure of multimodal images and construct the links between class relationships (same/different) and change labels (changed/unchanged) of pairwise superpixels, which are imaging modality-invariant. With these links, we formulate the MCD problem within a mathematical framework termed the locality-preserving energy model (LPEM), which is used to maintain the local consistency constraints embedded in the links: the structure consistency based on feature similarity and the label consistency based on spatial continuity. Because the foundation of LPEM, i.e., the links, is intuitively explainable and universal, the proposed method is very robust across different MCD situations. Noteworthy, LPEM is built directly on the label of each superpixel, so it is a paradigm that outputs the change map (CM) directly without the need to generate intermediate difference image (DI) as most previous algorithms have done. Experiments on different real datasets demonstrate the effectiveness of the proposed method. Source code of the proposed method is made available at https://github.com/yulisun/LPEM.
Collapse
|
2
|
Li Y, Zhang L, Shao L. LR Aerial Photo Categorization by Cross-Resolution Perceptual Knowledge Propagation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:3384-3395. [PMID: 38252579 DOI: 10.1109/tnnls.2024.3349515] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/24/2024]
Abstract
There are hundreds of high- and low-altitude earth observation satellites that asynchronously capture massive-scale aerial photographs every day. Generally, high-altitude satellites take low-resolution (LR) aerial pictures, each covering a considerably large area. In contrast, low-altitude satellites capture high-resolution (HR) aerial photos, each depicting a relatively small area. Accurately discovering the semantics of LR aerial photos is an indispensable technique in computer vision. Nevertheless, it is also a challenging task due to: 1) the difficulty to characterize human hierarchical visual perception and 2) the intolerable human resources to label sufficient training data. To handle these problems, a novel cross-resolution perceptual knowledge propagation (CPKP) framework is proposed, focusing on adapting the visual perceptual experiences deeply learned from HR aerial photos to categorize LR ones. Specifically, by mimicking the human vision system, a novel low-rank model is designed to decompose each LR aerial photo into multiple visually/semantically salient foreground regions coupled with the background nonsalient regions. This model can: 1) produce a gaze-shifting path (GSP) simulating human gaze behavior and 2) engineer the deep feature for each GSP. Afterward, a kernel-induced feature selection (FS) algorithm is formulated to obtain a succinct set of deep GSP features discriminative across LR and HR aerial photos. Based on the selected features, the labels from LR and HR aerial photos are collaboratively utilized to train a linear classifier for categorizing LR ones. It is worth emphasizing that, such a CPKP mechanism can effectively optimize the linear classifier training, as labels of HR aerial photos are acquired more conveniently in practice. Comprehensive visualization results and comparative study have validated the superiority of our approach.
Collapse
|
3
|
Qu J, Dong W, Yang Y, Zhang T, Li Y, Du Q. Cycle-Refined Multidecision Joint Alignment Network for Unsupervised Domain Adaptive Hyperspectral Change Detection. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:2634-2647. [PMID: 38170657 DOI: 10.1109/tnnls.2023.3347301] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/05/2024]
Abstract
Hyperspectral change detection, which provides abundant information on land cover changes in the Earth's surface, has become one of the most crucial tasks in remote sensing. Recently, deep-learning-based change detection methods have shown remarkable performance, but the acquirement of labeled data is extremely expensive and time-consuming. It is intuitive to learn changes from the scene with sufficient labeled data and adapting them into an unlabeled new scene. However, the nonnegligible domain shift between different scenes leads to inevitable performance degradation. In this article, a cycle-refined multidecision joint alignment network (CMJAN) is proposed for unsupervised domain adaptive hyperspectral change detection, which realizes progressive alignment of the data distributions between the source and target domains with cycle-refined high-confidence labeled samples. There are two key characteristics: 1) progressively mitigate the distribution discrepancy to learn domain-invariant difference feature representation and 2) update the high-confidence training samples of the target domain in a cycle manner. The benefit is that the domain shift between the source and target domains is progressively alleviated to promote change detection performance on the target domain in an unsupervised manner. Experimental results on different datasets demonstrate that the proposed method can achieve better performance than the state-of-the-art change detection methods.
Collapse
|
4
|
Kang X, Duan P, Li J, Li S. Efficient Swin Transformer for Remote Sensing Image Super-Resolution. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:6367-6379. [PMID: 39504286 DOI: 10.1109/tip.2024.3489228] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2024]
Abstract
Remote sensing super-resolution (SR) technique, which aims to generate high-resolution image with rich spatial details from its low-resolution counterpart, play a vital role in many applications. Recently, more and more studies attempt to explore the application of Transformer in remote sensing field. However, they suffer from the high computational burden and memory consumption for remote sensing super-resolution. In this paper, we propose an efficient Swin Transformer (ESTNet) via channel attention for SR of remote sensing images, which is composed of three components. First, a three-layer convolutional operation is utilized to extract shallow features of the input low-resolution image. Then, a residual group-wise attention module is proposed to extract the deep features, which contains an efficient channel attention block (ECAB) and a group-wise attention block (GAB). Finally, the extracted deep features are reconstructed to generate high-resolution remote sensing images. Extensive experimental results proclaim that the proposed ESTNet can obtain better super-resolution results with low computational burden. Compared to the recently proposed Transformer-based remote sensing super-resolution method, the number of parameters is reduced by 82.68% while the computational cost is reduced by 87.84%. The code of the proposed ESTNet will be available at https://github.com/PuhongDuan/ESTNet for reproducibility.
Collapse
|
5
|
Zhang M, Gao T, Gong M, Zhu S, Wu Y, Li H. Semisupervised Change Detection Based on Bihierarchical Feature Aggregation and Extraction Network. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:10488-10502. [PMID: 37022855 DOI: 10.1109/tnnls.2023.3242075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
With the rapid development of remote sensing (RS) technology, high-resolution RS image change detection (CD) has been widely used in many applications. Pixel-based CD techniques are maneuverable and widely used, but vulnerable to noise interference. Object-based CD techniques can effectively utilize the abundant spectrum, texture, shape, and spatial information but easy-to-ignore details of RS images. How to combine the advantages of pixel-based methods and object-based methods remains a challenging problem. Besides, although supervised methods have the capability to learn from data, the true labels representing changed information of RS images are often hard to obtain. To address these issues, this article proposes a novel semisupervised CD framework for high-resolution RS images, which employs small amounts of true labeled data and a lot of unlabeled data to train the CD network. A bihierarchical feature aggregation and extraction network (BFAEN) is designed to achieve the pixelwise together with objectwise feature concatenation feature representation for the comprehensive utilization of the two-level features. In order to alleviate the coarseness and insufficiency of labeled samples, a confident learning algorithm is used to eliminate noisy labels and a novel loss function is designed for training the model using true- and pseudo-labels in a semisupervised fashion. Experimental results on real datasets demonstrate the effectiveness and superiority of the proposed method.
Collapse
|
6
|
Li Z, Hu J, Wu K, Miao J, Zhao Z, Wu J. Local feature acquisition and global context understanding network for very high-resolution land cover classification. Sci Rep 2024; 14:12597. [PMID: 38824153 PMCID: PMC11144191 DOI: 10.1038/s41598-024-63363-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2024] [Accepted: 05/28/2024] [Indexed: 06/03/2024] Open
Abstract
Very high-resolution remote sensing images hold promising applications in ground observation tasks, paving the way for highly competitive solutions using image processing techniques for land cover classification. To address the challenges faced by convolutional neural network (CNNs) in exploring contextual information in remote sensing image land cover classification and the limitations of vision transformer (ViT) series in effectively capturing local details and spatial information, we propose a local feature acquisition and global context understanding network (LFAGCU). Specifically, we design a multidimensional and multichannel convolutional module to construct a local feature extractor aimed at capturing local information and spatial relationships within images. Simultaneously, we introduce a global feature learning module that utilizes multiple sets of multi-head attention mechanisms for modeling global semantic information, abstracting the overall feature representation of remote sensing images. Validation, comparative analyses, and ablation experiments conducted on three different scales of publicly available datasets demonstrate the effectiveness and generalization capability of the LFAGCU method. Results show its effectiveness in locating category attribute information related to remote sensing areas and its exceptional generalization capability. Code is available at https://github.com/lzp-lkd/LFAGCU .
Collapse
Affiliation(s)
- Zhengpeng Li
- School of Electronic and Information Engineering, University of Science and Technology Liaoning, Anshan, China
- Liaoning Province Key Laboratory of Intelligent Construction and Internet of Things Application Technologies, Anshan, China
| | - Jun Hu
- School of Electronic and Information Engineering, University of Science and Technology Liaoning, Anshan, China.
- Liaoning Province Key Laboratory of Intelligent Construction and Internet of Things Application Technologies, Anshan, China.
| | - Kunyang Wu
- College of Instrumentation and Electrical Engineering, Jilin University, Changchun, China
- National Geophysical Exploration Equipment Engineering Research Center, Jilin University, Changchun, China
- Key Laboratory of Geophysical Exploration Equipment Ministry of Education of China (Jilin University), Changchun, China
| | - Jiawei Miao
- School of Electronic and Information Engineering, University of Science and Technology Liaoning, Anshan, China
- Liaoning Province Key Laboratory of Intelligent Construction and Internet of Things Application Technologies, Anshan, China
| | - Zixue Zhao
- School of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, China
| | - Jiansheng Wu
- School of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, China
| |
Collapse
|
7
|
Gut D, Trombini M, Kucybała I, Krupa K, Rozynek M, Dellepiane S, Tabor Z, Wojciechowski W. Use of superpixels for improvement of inter-rater and intra-rater reliability during annotation of medical images. Med Image Anal 2024; 94:103141. [PMID: 38489896 DOI: 10.1016/j.media.2024.103141] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2022] [Revised: 01/29/2024] [Accepted: 03/05/2024] [Indexed: 03/17/2024]
Abstract
In the context of automatic medical image segmentation based on statistical learning, raters' variability of ground truth segmentations in training datasets is a widely recognized issue. Indeed, the reference information is provided by experts but bias due to their knowledge may affect the quality of the ground truth data, thus hindering creation of robust and reliable datasets employed in segmentation, classification or detection tasks. In such a framework, automatic medical image segmentation would significantly benefit from utilizing some form of presegmentation during training data preparation process, which could lower the impact of experts' knowledge and reduce time-consuming labeling efforts. The present manuscript proposes a superpixels-driven procedure for annotating medical images. Three different superpixeling methods with two different number of superpixels were evaluated on three different medical segmentation tasks and compared with manual annotations. Within the superpixels-based annotation procedure medical experts interactively select superpixels of interest, apply manual corrections, when necessary, and then the accuracy of the annotations, the time needed to prepare them, and the number of manual corrections are assessed. In this study, it is proven that the proposed procedure reduces inter- and intra-rater variability leading to more reliable annotations datasets which, in turn, may be beneficial for the development of more robust classification or segmentation models. In addition, the proposed approach reduces time needed to prepare the annotations.
Collapse
Affiliation(s)
- Daniel Gut
- Department of Biocybernetics and Biomedical Engineering, AGH University of Krakow, al. Mickiewicza 30, 30-059 Krakow, Poland.
| | - Marco Trombini
- Department of Electric, Electronic, and Telecommunication Engineering and Naval Architecture - DITEN, Università degli Studi di Genova, Via all'Opera Pia 11, 16145 Genoa, Italy
| | - Iwona Kucybała
- Department of Radiology, Jagiellonian University Medical College, ul. Kopernika 19, 31-501 Krakow, Poland
| | - Kamil Krupa
- Department of Radiology, Jagiellonian University Medical College, ul. Kopernika 19, 31-501 Krakow, Poland
| | - Miłosz Rozynek
- Department of Radiology, Jagiellonian University Medical College, ul. Kopernika 19, 31-501 Krakow, Poland
| | - Silvana Dellepiane
- Department of Electric, Electronic, and Telecommunication Engineering and Naval Architecture - DITEN, Università degli Studi di Genova, Via all'Opera Pia 11, 16145 Genoa, Italy
| | - Zbisław Tabor
- Department of Biocybernetics and Biomedical Engineering, AGH University of Krakow, al. Mickiewicza 30, 30-059 Krakow, Poland
| | - Wadim Wojciechowski
- Department of Radiology, Jagiellonian University Medical College, ul. Kopernika 19, 31-501 Krakow, Poland
| |
Collapse
|
8
|
Guo H, Su X, Wu C, Du B, Zhang L. SAAN: Similarity-Aware Attention Flow Network for Change Detection With VHR Remote Sensing Images. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:2599-2613. [PMID: 38427550 DOI: 10.1109/tip.2024.3349868] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/03/2024]
Abstract
Change detection (CD) is a fundamental and important task for monitoring the land surface dynamics in the earth observation field. Existing deep learning-based CD methods typically extract bi-temporal image features using a weight-sharing Siamese encoder network and identify change regions using a decoder network. These CD methods, however, still perform far from satisfactorily as we observe that 1) deep encoder layers focus on irrelevant background regions; and 2) the models' confidence in the change regions is inconsistent at different decoder stages. The first problem is because deep encoder layers cannot effectively learn from imbalanced change categories using the sole output supervision, while the second problem is attributed to the lack of explicit semantic consistency preservation. To address these issues, we design a novel similarity-aware attention flow network (SAAN). SAAN incorporates a similarity-guided attention flow module with deeply supervised similarity optimization to achieve effective change detection. Specifically, we counter the first issue by explicitly guiding deep encoder layers to discover semantic relations from bi-temporal input images using deeply supervised similarity optimization. The extracted features are optimized to be semantically similar in the unchanged regions and dissimilar in the changing regions. The second drawback can be alleviated by the proposed similarity-guided attention flow module, which incorporates similarity-guided attention modules and attention flow mechanisms to guide the model to focus on discriminative channels and regions. We evaluated the effectiveness and generalization ability of the proposed method by conducting experiments on a wide range of CD tasks. The experimental results demonstrate that our method achieves excellent performance on several CD tasks, with discriminative features and semantic consistency preserved.
Collapse
|
9
|
Okorie A, Kambhamettu C, Makrogiannnis S. Unsupervised learning of probabilistic subspaces for multi-spectral and multi-temporal image-based disaster mapping. MACHINE VISION AND APPLICATIONS 2023; 34:103. [PMID: 38586579 PMCID: PMC10997379 DOI: 10.1007/s00138-023-01451-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Revised: 07/24/2023] [Accepted: 08/15/2023] [Indexed: 04/09/2024]
Abstract
Accurate and timely identification of regions damaged by a natural disaster is critical for assessing the damages and reducing the human life cost. The increasing availability of satellite imagery and other remote sensing data has triggered research activities on development of algorithms for detection and monitoring of natural events. Here, we introduce an unsupervised subspace learning-based methodology that uses multi-temporal and multi-spectral satellite images to identify regions damaged by natural disasters. It first performs region delineation, matching, and fusion. Next, it applies subspace learning in the joint regional space to produce a change map. It identifies the damaged regions by estimating probabilistic subspace distances and rejecting the non-disaster changes. We evaluated the performance of our method on seven disaster datasets including four wildfire events, two flooding events, and a earthquake/tsunami event. We validated our results by calculating the dice similarity coefficient (DSC), and accuracy of classification between our disaster maps and ground-truth data. Our method produced average DSC values of 0.833 and 0.736, for wildfires and floods, respectively, and overall DSC of 0.855 for the tsunami event. The evaluation results support the applicability of our method to multiple types of natural disasters.
Collapse
Affiliation(s)
- Azubuike Okorie
- Division of Physics, Engineering, Mathematics, and Computer Sciences, Delaware State University, 1200 N. DuPont Hwy, Dover, DE 19901, USA
| | - Chandra Kambhamettu
- Department of Computer and Information Sciences, University of Delaware, 210 South College Avenue, Newark, DE 19716, USA
| | - Sokratis Makrogiannnis
- Division of Physics, Engineering, Mathematics, and Computer Sciences, Delaware State University, 1200 N. DuPont Hwy, Dover, DE 19901, USA
| |
Collapse
|
10
|
Sumbul G, Demir B. Generative Reasoning Integrated Label Noise Robust Deep Image Representation Learning. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:4529-4542. [PMID: 37440393 DOI: 10.1109/tip.2023.3293776] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/15/2023]
Abstract
The development of deep learning based image representation learning (IRL) methods has attracted great attention for various image understanding problems. Most of these methods require the availability of a set of high quantity and quality of annotated training images, which can be time-consuming, complex and costly to gather. To reduce labeling costs, crowdsourced data, automatic labeling procedures or citizen science projects can be considered. However, such approaches increase the risk of including label noise in training data. It may result in overfitting on noisy labels when discriminative reasoning is employed as in most of the existing methods. This leads to sub-optimal learning procedures, and thus inaccurate characterization of images. To address this issue, in this paper, we introduce a generative reasoning integrated label noise robust deep representation learning (GRID) approach. The proposed GRID approach aims to model the complementary characteristics of discriminative and generative reasoning for IRL under noisy labels. To this end, we first integrate generative reasoning into discriminative reasoning through a supervised variational autoencoder. This allows the proposed GRID approach to automatically detect training samples with noisy labels. Then, through our label noise robust hybrid representation learning strategy, GRID adjusts the whole learning procedure for IRL of these samples through generative reasoning and that of the other samples through discriminative reasoning. Our approach learns discriminative image representations while preventing interference of noisy labels during training independently from the IRL method being selected. Thus, unlike the existing label noise robust methods, GRID does not depend on the type of annotation, label noise, neural network architecture, loss function or learning task, and thus can be directly utilized for various image understanding problems. Experimental results show the effectiveness of the proposed GRID approach compared to the state-of-the-art methods. The code of the proposed approach is publicly available at https://github.com/gencersumbul/GRID.
Collapse
|
11
|
Liu W, Jin F. Electrochemical Intelligent Recognition of Mineral Materials Based on Superpixel Image Segmentation. Int J Anal Chem 2022; 2022:6755771. [PMID: 35756146 PMCID: PMC9217615 DOI: 10.1155/2022/6755771] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2022] [Revised: 06/02/2022] [Accepted: 06/04/2022] [Indexed: 11/19/2022] Open
Abstract
In order to study the needs of identifying rock thin-section samples by manual observation in the field of geology, a method of electrochemical intelligent recognition of mineral materials based on superpixel image segmentation is proposed. The image histogram of this method can be used to represent the distribution of each pixel value of the image. This interval is consistent with the number of pixels in the method. And using the experiment, the CPU used in the experiment is Intel® Core™ i7-8700 3.2 GHz, the memory is 16 GB, and the GPU is NVIDIA GeForce GT × 1080 Ti, which ensures the accuracy of the experiment. Based on all the experimental results, it can be seen that after the two-stage processing of the designed superpixel algorithm and the region merging algorithm, the final sandstone slice image segmentation results are close to the results of manual labeling, which is helpful for the subsequent research on sandstone component identification. The feasibility of this method was verified.
Collapse
Affiliation(s)
- Weiping Liu
- Department of Fundamental Subjects, Wuchang Shouyi University, Wuhan 430064, China
| | - Fangzhou Jin
- Department of Fundamental Subjects, Wuchang Shouyi University, Wuhan 430064, China
| |
Collapse
|
12
|
MSGATN: A Superpixel-Based Multi-Scale Siamese Graph Attention Network for Change Detection in Remote Sensing Images. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12105158] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/04/2022]
Abstract
With the rapid development of Earth observation technology, how to effectively and efficiently detect changes in multi-temporal images has become an important but challenging problem. Relying on the advantages of high performance and robustness, object-based change detection (CD) has become increasingly popular. By analyzing the similarity of local pixels, object-based CD aggregates similar pixels into one object and takes it as the basic processing unit. However, object-based approaches often have difficulty capturing discriminative features, as irregular objects make processing difficult. To address this problem, in this paper, we propose a novel superpixel-based multi-scale Siamese graph attention network (MSGATN) which can process unstructured data natively and extract valuable features. First, a difference image (DI) is generated by Euclidean distance between bitemporal images. Second, superpixel segmentation is employed based on DI to divide each image into many homogeneous regions. Then, these superpixels are used to model the problem by graph theory to construct a series of nodes with the adjacency between them. Subsequently, the multi-scale neighborhood features of the nodes are extracted through applying a graph convolutional network and concatenated by an attention mechanism. Finally, the binary change map can be obtained by classifying each node by some fully connected layers. The novel features of MSGATN can be summarized as follows: (1) Training in multi-scale constructed graphs improves the recognition over changed land cover of varied sizes and shapes. (2) Spectral and spatial self-attention mechanisms are exploited for a better change detection performance. The experimental results on several real datasets show the effectiveness and superiority of the proposed method. In addition, compared to other recent methods, the proposed can demonstrate very high processing efficiency and greatly reduce the dependence on labeled training samples in a semisupervised training fashion.
Collapse
|
13
|
SMD-Net: Siamese Multi-Scale Difference-Enhancement Network for Change Detection in Remote Sensing. REMOTE SENSING 2022. [DOI: 10.3390/rs14071580] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
Change detection, as an important task of remote sensing image processing, has a wide range of applications in many aspects such as land use and natural disaster assessment. Recent change detection methods have achieved good results. However, due to the environmental difference between the bi-temporal images and the complicated imaging condition, there are usually problems such as missing small objects, incomplete objects, and rough edges in the change detection results. The existing change detection methods usually lack attention in these areas. In this paper, we propose a Siamese change detection method, named SMD-Net, for bi-temporal remote sensing change detection. The proposed model uses multi-scale difference maps to enhances the information of the changed areas step by step in order to have better change detection results. Furthermore, we propose a Siamese residual multi-kernel pooling module (SRMP) for high-level features to enhance the high-level change information of the model. For the low-level features of multiple skip connections, we propose a feature difference module (FDM) that uses feature difference to fully extract the change information and help the model generate more accurate details. The experimental results of our method on three public datasets show that compared with other benchmark methods, our network comprises better effectiveness and has a better trade-off between accuracy and calculation cost.
Collapse
|