1
|
Sevugan P, Rudhrakoti V, Kim TH, Gunasekaran M, Purushotham S, Chinthaginjala R, Ahmad I, Kumar A.. Class-aware feature attention-based semantic segmentation on hyperspectral images. PLoS One 2025; 20:e0309997. [PMID: 39903744 PMCID: PMC11793730 DOI: 10.1371/journal.pone.0309997] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Accepted: 08/23/2024] [Indexed: 02/06/2025] Open
Abstract
This research explores an innovative approach to segment hyperspectral images. Aclass-aware feature-based attention approach is combined with an enhanced attention-based network, FAttNet is proposed to segment the hyperspectral images semantically. It is introduced to address challenges associated with inaccurate edge segmentation, diverse forms of target inconsistency, and suboptimal predictive efficacy encountered in traditional segmentation networks when applied to semantic segmentation tasks in hyperspectral images. First, the class-aware feature attention procedure is used to improve the extraction and processing of distinct types of semantic information. Subsequently, the spatial attention pyramid is employed in a parallel fashion to improve the correlation between spaces and extract context information from images at different scales. Finally, the segmentation results are refined using the encoder-decoder structure. It enhances precision in delineating distinct land cover patterns. The findings from the experiments demonstrate that FAttNet exhibits superior performance compared to established semantic segmentation networks commonly used. Specifically, on the GaoFen image dataset, FAttNet achieves a higher mean intersection over union (MIoU) of 77.03% and a segmentation accuracy of 87.26% surpassing the performance of the existing network.
Collapse
Affiliation(s)
- Prabu Sevugan
- Department of Banking Technology, Pondicherry University (A Central University), Puducherry, India
| | | | - Tai-hoon Kim
- School of Electrical and Computer Engineering, Yeosu Campus, Chonnam National University, Yeosu-si, Jeollanam-do, Republic of Korea
| | - Megala Gunasekaran
- School of Computer Science and Engineering, Vellore Institute of Technology at Vellore, Vellore, India
| | - Swarnalatha Purushotham
- School of Computer Science and Engineering, Vellore Institute of Technology at Vellore, Vellore, India
| | | | - Irfan Ahmad
- Department of Clinical Laboratory Science, College of Applied Medical Sciences, King Khalid University, Abha, Saudi Arabia
| | - Kumar A.
- Data Science Research Laboratory, BlueCrest University, Monrovia, Liberia
| |
Collapse
|
2
|
Cardone B, Di Martino F, Miraglia V. A Novel Fuzzy-Based Remote Sensing Image Segmentation Method. SENSORS (BASEL, SWITZERLAND) 2023; 23:9641. [PMID: 38139487 PMCID: PMC10747474 DOI: 10.3390/s23249641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Revised: 11/24/2023] [Accepted: 12/04/2023] [Indexed: 12/24/2023]
Abstract
Image segmentation is a well-known image processing task that consists of partitioning an image into homogeneous areas. It is applied to remotely sensed imagery for many problems such as land use classification and landscape changes. Recently, several hybrid remote sensing image segmentation techniques have been proposed that include metaheuristic approaches in order to increase the segmentation accuracy; however, the critical point of these approaches is the high computational complexity, which affects time and memory consumption. In order to overcome this criticality, we propose a fuzzy-based image segmentation framework implemented in a GIS-based platform for remotely sensed images; furthermore, the proposed model allows us to evaluate the reliability of the segmentation. The Fast Generalized Fuzzy c-means algorithm is implemented to segment images in order to detect local spatial relations between pixels and the Triple Center Relation validity index is used to find the optimal number of clusters. The framework elaborates the composite index to be analyzed starting by multiband remotely sensed images. For each cluster, a segmented image is obtained in which the pixel value represents, transformed into gray levels, the graph belonging to the cluster. A final thematic map is built in which the pixels are classified based on the assignment to the cluster to which they belong with the highest membership degree. In addition, the reliability of the classification is estimated by associating each class with the average of the membership degrees of the pixels assigned to it. The method was tested in the study area consisting of the south-western districts of the city of Naples (Italy) for the segmentation of composite indices maps determined by multiband remote sensing images. The segmentation results are consistent with the segmentations of the study area by morphological and urban characteristics, carried out by domain experts. The high computational speed of the proposed image segmentation method allows it to be applied to massive high-resolution remote sensing images.
Collapse
Affiliation(s)
- Barbara Cardone
- Department of Architecture, University of Naples Federico II, Via Toledo 402, 80134 Naples, Italy; (B.C.); (V.M.)
| | - Ferdinando Di Martino
- Department of Architecture, University of Naples Federico II, Via Toledo 402, 80134 Naples, Italy; (B.C.); (V.M.)
- Center for Interdepartmental Research “Alberto Calza Bini”, University of Naples Federico II, Via Toledo 402, 80134 Naples, Italy
| | - Vittorio Miraglia
- Department of Architecture, University of Naples Federico II, Via Toledo 402, 80134 Naples, Italy; (B.C.); (V.M.)
| |
Collapse
|
3
|
Context-content collaborative network for building extraction from high-resolution imagery. Knowl Based Syst 2023. [DOI: 10.1016/j.knosys.2023.110283] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
|
4
|
Zhao J, Zhang D, Shi B, Zhou Y, Chen J, Yao R, Xue Y. Multi-source collaborative enhanced for remote sensing images semantic segmentation. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.04.045] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
5
|
Li Y, Ouyang S, Zhang Y. Combining deep learning and ontology reasoning for remote sensing image semantic segmentation. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2022.108469] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
6
|
Sun Y, Ye Y, Li X, Feng S, Zhang B, Kang J, Dai K. Unsupervised deep hashing through learning soft pseudo label for remote sensing image retrieval. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2021.107807] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
7
|
Wójcicka A, Walusiak Ł, Mroczka K, Jaworek-Korjakowska JK, Oprzędkiewicz K, Wrobel Z. The Object Segmentation from the Microstructure of a FSW Dissimilar Weld. MATERIALS 2022; 15:ma15031129. [PMID: 35161074 PMCID: PMC8839914 DOI: 10.3390/ma15031129] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/28/2021] [Revised: 01/25/2022] [Accepted: 01/28/2022] [Indexed: 02/01/2023]
Abstract
Friction stir welding (FSW) is an environmentally friendly, solid-state welding technique. In this research work, we analyze the microstructure of a new type of FSW weld applying a two- stage framework based on image processing algorithms containing a segmentation step and microstructure analysis of objects occurring in different layers. A dual-speed tool as used to prepare the tested weld. In this paper, we present the segmentation method for recognizing areas containing particles forming bands in the microstructure of a dissimilar weld of aluminum alloys made by FSW technology. A digital analysis was performed on the images obtained using an Olympus GX51 light microscope. The image analysis process consisted of basic segmentation methods in conjunction with domain knowledge and object detection located in different layers of a weld using morphological operations and point transformations. These methods proved to be effective in the analysis of the microstructure images corrupted by noise. The segmentation parts as well as single objects were separated enough to analyze the distribution on different layers of the specimen and the variability of shape and size of the underlying microstructures, which was not possible without computer vision support.
Collapse
Affiliation(s)
- Anna Wójcicka
- Department of Automatic Control and Robotics, AGH University of Science and Technology, 30-059 Cracow, Poland; (J.K.J.-K.); (K.O.)
- Correspondence:
| | - Łukasz Walusiak
- Faculty of Architecture, Civil Engineering and Applied Arts, University of Technology, Rolna 43, 40-555 Katowice, Poland;
| | - Krzysztof Mroczka
- Faculty of Materials Engineering and Physics, Cracow University of Technology, 31-864 Cracow, Poland;
| | | | - Krzysztof Oprzędkiewicz
- Department of Automatic Control and Robotics, AGH University of Science and Technology, 30-059 Cracow, Poland; (J.K.J.-K.); (K.O.)
| | - Zygmunt Wrobel
- Institute of Biomedical Engineering, Faculty of Science and Technology, University of Silesia in Katowice, 41-205 Sosnowiec, Poland;
| |
Collapse
|
8
|
|
9
|
Abstract
The precise segmentation of forest areas is essential for monitoring tasks related to forest exploration, extraction, and statistics. However, the effective and accurate segmentation of forest images will be affected by factors such as blurring and discontinuity of forest boundaries. Therefore, a Pyramid Feature Extraction-UNet network (PFE-UNet) based on traditional UNet is proposed to be applied to end-to-end forest image segmentation. Among them, the Pyramid Feature Extraction module (PFE) is introduced in the network transition layer, which obtains multi-scale forest image information through different receptive fields. The spatial attention module (SA) and the channel-wise attention module (CA) are applied to low-level feature maps and PFE feature maps, respectively, to highlight specific segmentation task features while fusing context information and suppressing irrelevant regions. The standard convolution block is replaced by a novel depthwise separable convolutional unit (DSC Unit), which not only reduces the computational cost but also prevents overfitting. This paper presents an extensive evaluation with the DeepGlobe dataset and a comparative analysis with several state-of-the-art networks. The experimental results show that the PFE-UNet network obtains an accuracy of 94.23% in handling the real-time forest image segmentation, which is significantly higher than other advanced networks. This means that the proposed PFE-UNet also provides a valuable reference for the precise segmentation of forest images.
Collapse
|
10
|
Combining Deep Semantic Segmentation Network and Graph Convolutional Neural Network for Semantic Segmentation of Remote Sensing Imagery. REMOTE SENSING 2020. [DOI: 10.3390/rs13010119] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Although the deep semantic segmentation network (DSSN) has been widely used in remote sensing (RS) image semantic segmentation, it still does not fully mind the spatial relationship cues between objects when extracting deep visual features through convolutional filters and pooling layers. In fact, the spatial distribution between objects from different classes has a strong correlation characteristic. For example, buildings tend to be close to roads. In view of the strong appearance extraction ability of DSSN and the powerful topological relationship modeling capability of the graph convolutional neural network (GCN), a DSSN-GCN framework, which combines the advantages of DSSN and GCN, is proposed in this paper for RS image semantic segmentation. To lift the appearance extraction ability, this paper proposes a new DSSN called the attention residual U-shaped network (AttResUNet), which leverages residual blocks to encode feature maps and the attention module to refine the features. As far as GCN, the graph is built, where graph nodes are denoted by the superpixels and the graph weight is calculated by considering the spectral information and spatial information of the nodes. The AttResUNet is trained to extract the high-level features to initialize the graph nodes. Then the GCN combines features and spatial relationships between nodes to conduct classification. It is worth noting that the usage of spatial relationship knowledge boosts the performance and robustness of the classification module. In addition, benefiting from modeling GCN on the superpixel level, the boundaries of objects are restored to a certain extent and there are less pixel-level noises in the final classification result. Extensive experiments on two publicly open datasets show that DSSN-GCN model outperforms the competitive baseline (i.e., the DSSN model) and the DSSN-GCN when adopting AttResUNet achieves the best performance, which demonstrates the advance of our method.
Collapse
|
11
|
Shen X, Liu B, Zhou Y, Zhao J, Liu M. Remote sensing image captioning via Variational Autoencoder and Reinforcement Learning. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2020.105920] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
|
12
|
Dong Z, Du X, Liu Y. Automatic segmentation of left ventricle using parallel end–end deep convolutional neural networks framework. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2020.106210] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/08/2023]
|
13
|
Gómez-Ríos A, Tabik S, Luengo J, Shihavuddin A, Herrera F. Coral species identification with texture or structure images using a two-level classifier based on Convolutional Neural Networks. Knowl Based Syst 2019. [DOI: 10.1016/j.knosys.2019.104891] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
14
|
Liping C, Saeed S, Yujun S. Image classification based on the linear unmixing and GEOBIA. ENVIRONMENTAL MONITORING AND ASSESSMENT 2019; 191:649. [PMID: 31624914 DOI: 10.1007/s10661-019-7837-x] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/16/2019] [Accepted: 09/24/2019] [Indexed: 06/10/2023]
Abstract
Geographic Object-Based Image Analysis and linear unmixing are common methods in image classification. The purpose of this study is to analyze the classification efficiency by integrating these two methods in the mountain area. This research selected Jiangle County, Fujian, as a study area. Two Landsat8 OLI images, which covered the county, were used. Linear spectral mixture model, multi-scale segmentation, and decision tree were applied in the classification. After image preprocessing, linear spectral mixture model was used to unmix the image into three fraction images-vegetation, shade, and soil. The principal component analysis and tasseled cap transformation were used to derived three principal components and the brightness, wetness, and greenness. Multi-scale segmentation is applied by eCognition. Under scale 40, the image was divided into vegetation and non-vegetation area, then under scale 20, the vegetation area was divided into different types by integrating the fraction with different methods. The accuracy assessment of the classification map was done using the forestry resource survey and the high-resolution image of Google Earth. This study indicated that the unmixed bands could improve the classification accuracy. The overall classification accuracy was 92.40% with a Kappa coefficient of 0.9032. Therefore, there is a conclusion that this approach is an efficient way to classify different plantation.
Collapse
Affiliation(s)
- Chen Liping
- State Forestry Administration Key Laboratory of Forest Resources & Environmental Management, Beijing Forestry University, Beijing, 100083, China
| | - Sajjad Saeed
- State Forestry Administration Key Laboratory of Forest Resources & Environmental Management, Beijing Forestry University, Beijing, 100083, China
| | - Sun Yujun
- State Forestry Administration Key Laboratory of Forest Resources & Environmental Management, Beijing Forestry University, Beijing, 100083, China.
| |
Collapse
|
15
|
Convolutional Neural Network-Based Remote Sensing Images Segmentation Method for Extracting Winter Wheat Spatial Distribution. APPLIED SCIENCES-BASEL 2018. [DOI: 10.3390/app8101981] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
When extracting winter wheat spatial distribution by using convolutional neural network (CNN) from Gaofen-2 (GF-2) remote sensing images, accurate identification of edge pixel is the key to improving the result accuracy. In this paper, an approach for extracting accurate winter wheat spatial distribution based on CNN is proposed. A hybrid structure convolutional neural network (HSCNN) was first constructed, which consists of two independent sub-networks of different depths. The deeper sub-network was used to extract the pixels present in the interior of the winter wheat field, whereas the shallower sub-network extracts the pixels at the edge of the field. The model was trained by classification-based learning and used in image segmentation for obtaining the distribution of winter wheat. Experiments were performed on 39 GF-2 images of Shandong province captured during 2017–2018, with SegNet and DeepLab as comparison models. As shown by the results, the average accuracy of SegNet, DeepLab, and HSCNN was 0.765, 0.853, and 0.912, respectively. HSCNN was equally as accurate as DeepLab and superior to SegNet for identifying interior pixels, and its identification of the edge pixels was significantly better than the two comparison models, which showed the superiority of HSCNN in the identification of winter wheat spatial distribution.
Collapse
|
16
|
Lee SJ, Yun JP, Koo G, Kim SW. End-to-end recognition of slab identification numbers using a deep convolutional neural network. Knowl Based Syst 2017. [DOI: 10.1016/j.knosys.2017.06.017] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|