1
|
Lopes T, Cavaco R, Capela D, Dias F, Teixeira J, Monteiro CS, Lima A, Guimarães D, Jorge PAS, Silva NA. Improving LIBS-based mineral identification with Raman imaging and spectral knowledge distillation. Talanta 2025; 283:127110. [PMID: 39520923 DOI: 10.1016/j.talanta.2024.127110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2024] [Revised: 10/15/2024] [Accepted: 10/24/2024] [Indexed: 11/16/2024]
Abstract
Combining data from different sensing modalities has been a promising research topic for building better and more reliable data-driven models. In particular, it is known that multimodal spectral imaging can improve the analytical capabilities of standalone spectroscopy techniques through fusion, hyphenation, or knowledge distillation techniques. In this manuscript, we focus on the latter, exploring how one can increase the performance of a Laser-induced Breakdown Spectroscopy system for mineral classification problems using additional spectral imaging techniques. Specifically, focusing on a scenario where Raman spectroscopy delivers accurate mineral classification performance, we show how to deploy a knowledge distillation pipeline where Raman spectroscopy may act as an autonomous supervisor for LIBS. For a case study concerning a challenging Li-bearing mineral identification of spodumene and petalite, our results demonstrate the advantages of this method in improving the performance of a single-technique system. LIBS trained with labels obtained by Raman presents an enhanced classification performance. Furthermore, leveraging the interpretability of the model deployed, the workflow opens opportunities for the deployment of assisted feature discovery pipelines, which may impact future academic and industrial applications.
Collapse
Affiliation(s)
- Tomás Lopes
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal; Departamento de Física e Astronomia, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal.
| | - Rafael Cavaco
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal; Departamento de Física e Astronomia, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Diana Capela
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal; Departamento de Física e Astronomia, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Filipa Dias
- Departamento de Geociências, Ambiente e Ordenamento do Território, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Joana Teixeira
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal; Departamento de Física e Astronomia, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Catarina S Monteiro
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Alexandre Lima
- Departamento de Geociências, Ambiente e Ordenamento do Território, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Diana Guimarães
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Pedro A S Jorge
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal; Departamento de Física e Astronomia, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| | - Nuno A Silva
- Center for Applied Photonics, INESC TEC, Rua do Campo Alegre 687, Porto, 4169-007, Portugal; Departamento de Física e Astronomia, Faculdade de Ciências da Universidade do Porto, Rua do Campo Alegre 687, Porto, 4169-007, Portugal
| |
Collapse
|
2
|
Lopes T, Capela D, Guimarães D, Ferreira MFS, Jorge PAS, Silva NA. From sensor fusion to knowledge distillation in collaborative LIBS and hyperspectral imaging for mineral identification. Sci Rep 2024; 14:9123. [PMID: 38643168 PMCID: PMC11032373 DOI: 10.1038/s41598-024-59553-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 04/12/2024] [Indexed: 04/22/2024] Open
Abstract
Multimodal spectral imaging offers a unique approach to the enhancement of the analytical capabilities of standalone spectroscopy techniques by combining information gathered from distinct sources. In this manuscript, we explore such opportunities by focusing on two well-known spectral imaging techniques, namely laser-induced breakdown spectroscopy, and hyperspectral imaging, and explore the opportunities of collaborative sensing for a case study involving mineral identification. In specific, the work builds upon two distinct approaches: a traditional sensor fusion, where we strive to increase the information gathered by including information from the two modalities; and a knowledge distillation approach, where the Laser Induced Breakdown spectroscopy is used as an autonomous supervisor for hyperspectral imaging. Our results show the potential of both approaches in enhancing the performance over a single modality sensing system, highlighting, in particular, the advantages of the knowledge distillation framework in maximizing the potential benefits of using multiple techniques to build more interpretable models and paving for industrial applications.
Collapse
Affiliation(s)
- Tomás Lopes
- INESC TEC, Center for Applied Photonics, 4169-007, Porto, Portugal
- Departamento de Física, Faculdade de Ciências da Universidade do Porto, 4169-007, Porto, Portugal
| | - Diana Capela
- INESC TEC, Center for Applied Photonics, 4169-007, Porto, Portugal
- Departamento de Física, Faculdade de Ciências da Universidade do Porto, 4169-007, Porto, Portugal
| | - Diana Guimarães
- INESC TEC, Center for Applied Photonics, 4169-007, Porto, Portugal
| | - Miguel F S Ferreira
- INESC TEC, Center for Applied Photonics, 4169-007, Porto, Portugal
- Departamento de Física, Faculdade de Ciências da Universidade do Porto, 4169-007, Porto, Portugal
| | - Pedro A S Jorge
- INESC TEC, Center for Applied Photonics, 4169-007, Porto, Portugal
- Departamento de Física, Faculdade de Ciências da Universidade do Porto, 4169-007, Porto, Portugal
| | - Nuno A Silva
- INESC TEC, Center for Applied Photonics, 4169-007, Porto, Portugal.
| |
Collapse
|
3
|
Safari M, Fatemi A, Archambault L. MedFusionGAN: multimodal medical image fusion using an unsupervised deep generative adversarial network. BMC Med Imaging 2023; 23:203. [PMID: 38062431 PMCID: PMC10704723 DOI: 10.1186/s12880-023-01160-w] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2023] [Accepted: 11/23/2023] [Indexed: 12/18/2023] Open
Abstract
PURPOSE This study proposed an end-to-end unsupervised medical fusion generative adversarial network, MedFusionGAN, to fuse computed tomography (CT) and high-resolution isotropic 3D T1-Gd Magnetic resonance imaging (MRI) image sequences to generate an image with CT bone structure and MRI soft tissue contrast to improve target delineation and to reduce the radiotherapy planning time. METHODS We used a publicly available multicenter medical dataset (GLIS-RT, 230 patients) from the Cancer Imaging Archive. To improve the models generalization, we consider different imaging protocols and patients with various brain tumor types, including metastases. The proposed MedFusionGAN consisted of one generator network and one discriminator network trained in an adversarial scenario. Content, style, and L1 losses were used for training the generator to preserve the texture and structure information of the MRI and CT images. RESULTS The MedFusionGAN successfully generates fused images with MRI soft-tissue and CT bone contrast. The results of the MedFusionGAN were quantitatively and qualitatively compared with seven traditional and eight deep learning (DL) state-of-the-art methods. Qualitatively, our method fused the source images with the highest spatial resolution without adding the image artifacts. We reported nine quantitative metrics to quantify the preservation of structural similarity, contrast, distortion level, and image edges in fused images. Our method outperformed both traditional and DL methods on six out of nine metrics. And it got the second performance rank for three and two quantitative metrics when compared with traditional and DL methods, respectively. To compare soft-tissue contrast, intensity profile along tumor and tumor contours of the fusion methods were evaluated. MedFusionGAN provides a more consistent, better intensity profile, and a better segmentation performance. CONCLUSIONS The proposed end-to-end unsupervised method successfully fused MRI and CT images. The fused image could improve targets and OARs delineation, which is an important aspect of radiotherapy treatment planning.
Collapse
Affiliation(s)
- Mojtaba Safari
- Département de Physique, de génie Physique et d'Optique, et Centre de Recherche sur le Cancer, Université Laval, Québec City, QC, Canada.
- Service de Physique Médicale et Radioprotection, Centre Intégré de Cancérologie, CHU de Québec - Université Laval et Centre de recherche du CHU de Québec, Québec City, QC, Canada.
| | - Ali Fatemi
- Department of Physics, Jackson State University, Jackson, MS, USA
- Department of Radiation Oncology, Gamma Knife Center, Merit Health Central, Jackson, MS, USA
| | - Louis Archambault
- Département de Physique, de génie Physique et d'Optique, et Centre de Recherche sur le Cancer, Université Laval, Québec City, QC, Canada
- Service de Physique Médicale et Radioprotection, Centre Intégré de Cancérologie, CHU de Québec - Université Laval et Centre de recherche du CHU de Québec, Québec City, QC, Canada
| |
Collapse
|
4
|
Liu J, Liu P, Shi T, Ke M, Xiong K, Liu Y, Chen L, Zhang L, Liang X, Li H, Lu S, Lan X, Niu G, Zhang J, Fei P, Gao L, Tang J. Flexible and broadband colloidal quantum dots photodiode array for pixel-level X-ray to near-infrared image fusion. Nat Commun 2023; 14:5352. [PMID: 37660051 PMCID: PMC10475073 DOI: 10.1038/s41467-023-40620-3] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2022] [Accepted: 08/02/2023] [Indexed: 09/04/2023] Open
Abstract
Combining information from multispectral images into a fused image is informative and beneficial for human or machine perception. Currently, multiple photodetectors with different response bands are used, which require complicated algorithms and systems to solve the pixel and position mismatch problem. An ideal solution would be pixel-level multispectral image fusion, which involves multispectral image using the same photodetector and circumventing the mismatch problem. Here we presented the potential of pixel-level multispectral image fusion utilizing colloidal quantum dots photodiode array, with a broadband response range from X-ray to near infrared and excellent tolerance for bending and X-ray irradiation. The colloidal quantum dots photodiode array showed a specific detectivity exceeding 1012 Jones in visible and near infrared range and a favorable volume sensitivity of approximately 2 × 105 μC Gy-1 cm-3 for X-ray irradiation. To showcase the advantages of pixel-level multispectral image fusion, we imaged a capsule enfolding an iron wire and soft plastic, successfully revealing internal information through an X-ray to near infrared fused image.
Collapse
Affiliation(s)
- Jing Liu
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
- Optical Valley Laboratory, 430074, Wuhan, P. R. China
- Wenzhou Advanced Manufacturing Technology Research Institute of Huazhong University of Science and Technology, 225 Chaoyang New Street, 325105, Wenzhou, P. R. China
| | - Peilin Liu
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Tailong Shi
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Mo Ke
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Kao Xiong
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Yuxuan Liu
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Long Chen
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Linxiang Zhang
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Xinyi Liang
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Hao Li
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
| | - Shuaicheng Lu
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
- Wenzhou Advanced Manufacturing Technology Research Institute of Huazhong University of Science and Technology, 225 Chaoyang New Street, 325105, Wenzhou, P. R. China
| | - Xinzheng Lan
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
- Optical Valley Laboratory, 430074, Wuhan, P. R. China
| | - Guangda Niu
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
- Optical Valley Laboratory, 430074, Wuhan, P. R. China
| | - Jianbing Zhang
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
- Optical Valley Laboratory, 430074, Wuhan, P. R. China
- Wenzhou Advanced Manufacturing Technology Research Institute of Huazhong University of Science and Technology, 225 Chaoyang New Street, 325105, Wenzhou, P. R. China
| | - Peng Fei
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China
- Optical Valley Laboratory, 430074, Wuhan, P. R. China
| | - Liang Gao
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China.
- Optical Valley Laboratory, 430074, Wuhan, P. R. China.
- Wenzhou Advanced Manufacturing Technology Research Institute of Huazhong University of Science and Technology, 225 Chaoyang New Street, 325105, Wenzhou, P. R. China.
| | - Jiang Tang
- Wuhan National Laboratory for Optoelectronics and School of Optical and Electronic Information, Huazhong University of Science and Technology, 1037 Luoyu Road, 430074, Wuhan, P. R. China.
- Optical Valley Laboratory, 430074, Wuhan, P. R. China.
- National Engineering Research Center for Laser Processing, 1037 Luoyu Road, 430074, Wuhan, P. R. China.
| |
Collapse
|
5
|
Kurban R. Gaussian of Differences: A Simple and Efficient General Image Fusion Method. ENTROPY (BASEL, SWITZERLAND) 2023; 25:1215. [PMID: 37628245 PMCID: PMC10453154 DOI: 10.3390/e25081215] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/03/2023] [Revised: 08/08/2023] [Accepted: 08/14/2023] [Indexed: 08/27/2023]
Abstract
The separate analysis of images obtained from a single source using different camera settings or spectral bands, whether from one or more than one sensor, is quite difficult. To solve this problem, a single image containing all of the distinctive pieces of information in each source image is generally created by combining the images, a process called image fusion. In this paper, a simple and efficient, pixel-based image fusion method is proposed that relies on weighting the edge information associated with each pixel of all of the source images proportional to the distance from their neighbors by employing a Gaussian filter. The proposed method, Gaussian of differences (GD), was evaluated using multi-modal medical images, multi-sensor visible and infrared images, multi-focus images, and multi-exposure images, and was compared to existing state-of-the-art fusion methods by utilizing objective fusion quality metrics. The parameters of the GD method are further enhanced by employing the pattern search (PS) algorithm, resulting in an adaptive optimization strategy. Extensive experiments illustrated that the proposed GD fusion method ranked better on average than others in terms of objective quality metrics and CPU time consumption.
Collapse
Affiliation(s)
- Rifat Kurban
- Department of Computer Engineering, Abdullah Gul University, 38080 Kayseri, Turkey
| |
Collapse
|
6
|
Bocca P, Orellana A, Soria C, Carelli R. On field disease detection in olive tree with vision systems. ARRAY 2023. [DOI: 10.1016/j.array.2023.100286] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/04/2023] Open
|
7
|
Zhan W, Wang J, Jiang Y, Chen Y, Zheng T, Hong Y. Infrared and Visible Image Fusion for Highlighting Salient Targets in the Night Scene. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1759. [PMID: 36554164 PMCID: PMC9778389 DOI: 10.3390/e24121759] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/23/2022] [Revised: 11/26/2022] [Accepted: 11/29/2022] [Indexed: 06/17/2023]
Abstract
The goal of infrared and visible image fusion in the night scene is to generate a fused image containing salient targets and rich textural details. However, the existing image fusion methods fail to take the unevenness of nighttime luminance into account. To address the above issue, an infrared and visible image fusion method for highlighting salient targets in the night scene is proposed. First of all, a global attention module is designed, which rescales the weights of different channels after capturing global contextual information. Second, the loss function is divided into the foreground loss and the background loss, forcing the fused image to retain rich texture details while highlighting the salient targets. Finally, a luminance estimation function is introduced to obtain the trade-off control parameters of the foreground loss function based on the nighttime luminance. It can effectively highlight salient targets by retaining the foreground information from the source images. Compared with other advanced methods, the experimental results adequately demonstrate the excellent fusion performance and generalization of the proposed method.
Collapse
|
8
|
Texture and Materials Image Classification Based on Wavelet Pooling Layer in CNN. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12073592] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Convolutional Neural Networks (CNNs) have recently been proposed as a solution in texture and material classification in computer vision. However, inside CNNs, the internal layers of pooling often cause a loss of information and, therefore, is detrimental to learning the architecture. Moreover, when considering images with repetitive and essential patterns, the loss of this information affects the performance of subsequent stages, such as feature extraction and analysis. In this paper, to solve this problem, we propose a classification system with a new pooling method called Discrete Wavelet Transform Pooling (DWTP). This method is based on the image decomposition into sub-bands, in which the first level sub-band is considered as its output. The objective is to obtain approximation and detail information. As a result, this information can be concatenated in different combinations. In addition, wavelet pooling uses wavelets to reduce the size of the feature map. Combining these methods provides acceptable classification performance for three databases (CIFAR-10, DTD, and FMD). We argue that this helps eliminate overfitting and that the learning graphs reflect that the datasets show learning generalization. Therefore, our results indicate that our method based on wavelet analysis is feasible for texture and material classification. Moreover, in some cases, it outperforms traditional methods.
Collapse
|
9
|
Lakshmi A, Rajasekaran MP, Jeevitha S, Selvendran S. An Adaptive MRI-PET Image Fusion Model Based on Deep Residual Learning and Self-Adaptive Total Variation. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2022. [DOI: 10.1007/s13369-020-05201-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
10
|
Mostaar A, Amini N. Deep learning approach for fusion of magnetic resonance imaging-positron emission tomography image based on extract image features using pretrained network (VGG19). JOURNAL OF MEDICAL SIGNALS & SENSORS 2022; 12:25-31. [PMID: 35265462 PMCID: PMC8804594 DOI: 10.4103/jmss.jmss_80_20] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2020] [Revised: 01/01/2021] [Accepted: 06/14/2021] [Indexed: 11/17/2022]
Abstract
Background: The fusion of images is an interesting way to display the information of some different images in one image together. In this paper, we present a deep learning network approach for fusion of magnetic resonance imaging (MRI) and positron emission tomography (PET) images. Methods: We fused two MRI and PET images automatically with a pretrained convolutional neural network (CNN, VGG19). First, the PET image was converted from red-green-blue space to hue-saturation-intensity space to save the hue and saturation information. We started with extracting features from images by using a pretrained CNN. Then, we used the weights extracted from two MRI and PET images to construct a fused image. Fused image was constructed with multiplied weights to images. For solving the problem of reduced contrast, we added the constant coefficient of the original image to the final result. Finally, quantitative criteria (entropy, mutual information, discrepancy, and overall performance [OP]) were applied to evaluate the results of fusion. We compared the results of our method with the most widely used methods in the spatial and transform domain. Results: The quantitative measurement values we used were entropy, mutual information, discrepancy, and OP that were 3.0319, 2.3993, 3.8187, and 0.9899, respectively. The final results showed that our method based on quantitative assessments was the best and easiest way to fused images, especially in the spatial domain. Conclusion: It concluded that our method used for MRI-PET image fusion was more accurate.
Collapse
|
11
|
A. Alabdulkreem E, Sedik A, D. Algarni A, M. El Banby G, E. Abd El-Samie F, F. Soliman N. Enhanced Robotic Vision System Based on Deep Learning and Image Fusion. COMPUTERS, MATERIALS & CONTINUA 2022; 73:1845-1861. [DOI: 10.32604/cmc.2022.023905] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/26/2021] [Accepted: 03/30/2022] [Indexed: 09/02/2023]
|
12
|
Xu J, Lu K, Wang H. Attention fusion network for multi-spectral semantic segmentation. Pattern Recognit Lett 2021. [DOI: 10.1016/j.patrec.2021.03.015] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
13
|
A Generative Adversarial Network for Infrared and Visible Image Fusion Based on Semantic Segmentation. ENTROPY 2021; 23:e23030376. [PMID: 33801048 PMCID: PMC8004063 DOI: 10.3390/e23030376] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/19/2021] [Revised: 03/15/2021] [Accepted: 03/17/2021] [Indexed: 11/26/2022]
Abstract
This paper proposes a new generative adversarial network for infrared and visible image fusion based on semantic segmentation (SSGAN), which can consider not only the low-level features of infrared and visible images, but also the high-level semantic information. Source images can be divided into foregrounds and backgrounds by semantic masks. The generator with a dual-encoder-single-decoder framework is used to extract the feature of foregrounds and backgrounds by different encoder paths. Moreover, the discriminator’s input image is designed based on semantic segmentation, which is obtained by combining the foregrounds of the infrared images with the backgrounds of the visible images. Consequently, the prominence of thermal targets in the infrared images and texture details in the visible images can be preserved in the fused images simultaneously. Qualitative and quantitative experiments on publicly available datasets demonstrate that the proposed approach can significantly outperform the state-of-the-art methods.
Collapse
|