1
|
Li B, Zhang L, Liu J, Peng H, Wang Q, Liu J. Multi-focus image fusion with parameter adaptive dual channel dynamic threshold neural P systems. Neural Netw 2024; 179:106603. [PMID: 39146717 DOI: 10.1016/j.neunet.2024.106603] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Revised: 07/06/2024] [Accepted: 08/05/2024] [Indexed: 08/17/2024]
Abstract
Multi-focus image fusion (MFIF) is an important technique that aims to combine the focused regions of multiple source images into a fully clear image. Decision-map methods are widely used in MFIF to maximize the preservation of information from the source images. While many decision-map methods have been proposed, they often struggle with difficulties in determining focus and non-focus boundaries, further affecting the quality of the fused images. Dynamic threshold neural P (DTNP) systems are computational models inspired by biological spiking neurons, featuring dynamic threshold and spiking mechanisms to better distinguish focused and unfocused regions for decision map generation. However, original DTNP systems require manual parameter configuration and have only one stimulus. Therefore, they are not suitable to be used directly for generating high-precision decision maps. To overcome these limitations, we propose a variant called parameter adaptive dual channel DTNP (PADCDTNP) systems. Inspired by the spiking mechanisms of PADCDTNP systems, we further develop a new MFIF method. As a new neural model, PADCDTNP systems adaptively estimate parameters according to multiple external inputs to produce decision maps with robust boundaries, resulting in high-quality fusion results. Comprehensive experiments on the Lytro/MFFW/MFI-WHU dataset show that our method achieves advanced performance and yields comparable results to the fourteen representative MFIF methods. In addition, compared to the standard DTNP systems, PADCDTNP systems improve the fusion performance and fusion efficiency on the three datasets by 5.69% and 86.03%, respectively. The codes for both the proposed method and the comparison methods are released at https://github.com/MorvanLi/MFIF-PADCDTNP.
Collapse
Affiliation(s)
- Bo Li
- School of Computer Science and Technology, Xi'an Jiaotong University, Xi'an, Shaanxi, 710049, China; Ministry of Education Key Laboratory of Intelligent Networks and Network Security, Xi'an Jiaotong University, Xi'an, 710049, China
| | - Lingling Zhang
- School of Computer Science and Technology, Xi'an Jiaotong University, Xi'an, Shaanxi, 710049, China; Ministry of Education Key Laboratory of Intelligent Networks and Network Security, Xi'an Jiaotong University, Xi'an, 710049, China.
| | - Jun Liu
- School of Computer Science and Technology, Xi'an Jiaotong University, Xi'an, Shaanxi, 710049, China; Shaanxi Province Key Laboratory of Big Data Knowledge Engineering, Xi'an Jiaotong University, Xi'an, 710049, China
| | - Hong Peng
- School of Computer and Software Engineering, Xihua University, Chengdu, Sichuan, 610039, China
| | | | - Jiaqi Liu
- Henan University of Chinese Medicine, Henan, 450046, China
| |
Collapse
|
2
|
Aymaz S, Köse C, Aymaz Ş. A novel approach with the dynamic decision mechanism (DDM) in multi-focus image fusion. MULTIMEDIA TOOLS AND APPLICATIONS 2023; 82:1821-1871. [DOI: 10.1007/s11042-022-13323-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/17/2021] [Revised: 05/16/2022] [Accepted: 05/30/2022] [Indexed: 10/04/2024]
|
3
|
Liu S, Peng W, Jiang W, Yang Y, Zhao J, Su Y. Multi-focus image fusion dataset and algorithm test in real environment. Front Neurorobot 2022; 16:1024742. [PMID: 36329789 PMCID: PMC9623155 DOI: 10.3389/fnbot.2022.1024742] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2022] [Accepted: 09/09/2022] [Indexed: 11/30/2022] Open
Affiliation(s)
- Shuaiqi Liu
- College of Electronic and Information Engineering, Hebei University, Baoding, China
- Machine Vision Technological Innovation Center of Hebei, Baoding, China
- National Laboratory of Pattern Recognition (NLPR), Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Weijian Peng
- College of Electronic and Information Engineering, Hebei University, Baoding, China
- Machine Vision Technological Innovation Center of Hebei, Baoding, China
| | - Wenjing Jiang
- College of Electronic and Information Engineering, Hebei University, Baoding, China
- Machine Vision Technological Innovation Center of Hebei, Baoding, China
| | - Yang Yang
- College of Electronic and Information Engineering, Hebei University, Baoding, China
- Machine Vision Technological Innovation Center of Hebei, Baoding, China
| | - Jie Zhao
- College of Electronic and Information Engineering, Hebei University, Baoding, China
- Machine Vision Technological Innovation Center of Hebei, Baoding, China
| | - Yonggang Su
- College of Electronic and Information Engineering, Hebei University, Baoding, China
- Machine Vision Technological Innovation Center of Hebei, Baoding, China
- *Correspondence: Yonggang Su
| |
Collapse
|
4
|
Conditional Random Field-Guided Multi-Focus Image Fusion. J Imaging 2022; 8:jimaging8090240. [PMID: 36135406 PMCID: PMC9505971 DOI: 10.3390/jimaging8090240] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2022] [Revised: 08/21/2022] [Accepted: 09/02/2022] [Indexed: 11/16/2022] Open
Abstract
Multi-Focus image fusion is of great importance in order to cope with the limited Depth-of-Field of optical lenses. Since input images contain noise, multi-focus image fusion methods that support denoising are important. Transform-domain methods have been applied to image fusion, however, they are likely to produce artifacts. In order to cope with these issues, we introduce the Conditional Random Field (CRF) CRF-Guided fusion method. A novel Edge Aware Centering method is proposed and employed to extract the low and high frequencies of the input images. The Independent Component Analysis—ICA transform is applied to high-frequency components and a Conditional Random Field (CRF) model is created from the low frequency and the transform coefficients. The CRF model is solved efficiently with the α-expansion method. The estimated labels are used to guide the fusion of the low-frequency components and the transform coefficients. Inverse ICA is then applied to the fused transform coefficients. Finally, the fused image is the addition of the fused low frequency and the fused high frequency. CRF-Guided fusion does not introduce artifacts during fusion and supports image denoising during fusion by applying transform domain coefficient shrinkage. Quantitative and qualitative evaluation demonstrate the superior performance of CRF-Guided fusion compared to state-of-the-art multi-focus image fusion methods.
Collapse
|
5
|
Zhang X. Deep Learning-Based Multi-Focus Image Fusion: A Survey and a Comparative Study. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:4819-4838. [PMID: 33974542 DOI: 10.1109/tpami.2021.3078906] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Multi-focus image fusion (MFIF) is an important area in image processing. Since 2017, deep learning has been introduced to the field of MFIF and various methods have been proposed. However, there is a lack of survey papers that discuss deep learning-based MFIF methods in detail. In this study, we fill this gap by giving a detailed survey on deep learning-based MFIF algorithms, including methods, datasets and evaluation metrics. To the best of our knowledge, this is the first survey paper that focuses on deep learning-based approaches in the field of MFIF. Besides, extensive experiments have been conducted to compare the performance of deep learning-based MFIF algorithms with conventional MFIF approaches. By analyzing qualitative and quantitative results, we give some observations on the current status of MFIF and discuss some future prospects of this field.
Collapse
|
6
|
Wang Z, Li X, Duan H, Zhang X. A Self-Supervised Residual Feature Learning Model for Multifocus Image Fusion. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:4527-4542. [PMID: 35737635 DOI: 10.1109/tip.2022.3184250] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Multi-focus image fusion (MFIF) attempts to achieve an "all-focused" image from multiple source images with the same scene but different focused objects. Given the lack of multi-focus image sets for network training, we propose a self-supervised residual feature learning model in this paper. The model consists of a feature extraction network and a fusion module. We select image super-resolution as a pretext task in the MFIF field, which is supported by a new residual gradient prior discovered by our theoretical study for low- and high-resolution (LR-HR) image pairs, as well as for multi-focus images. In the pretext task, our network's training set is LR-HR image pairs generated from natural images, and HR images can be regarded as pseudo-labels of LR images. In the fusion task, the trained network extracts residual features of multi-focus images firstly. Secondly, the fusion module, consisting of an activity level measurement and a new boundary refinement method, is leveraged for the features to generated decision maps. Experimental results, both subjective evaluations and objective evaluations, demonstrate that our approach outperforms other state-of-the-art fusion algorithms.
Collapse
|
7
|
Ilyas A, Farid MS, Khan MH, Grzegorzek M. Exploiting Superpixels for Multi-Focus Image Fusion. ENTROPY (BASEL, SWITZERLAND) 2021; 23:247. [PMID: 33670018 PMCID: PMC7926613 DOI: 10.3390/e23020247] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/29/2021] [Revised: 02/15/2021] [Accepted: 02/17/2021] [Indexed: 12/03/2022]
Abstract
Multi-focus image fusion is the process of combining focused regions of two or more images to obtain a single all-in-focus image. It is an important research area because a fused image is of high quality and contains more details than the source images. This makes it useful for numerous applications in image enhancement, remote sensing, object recognition, medical imaging, etc. This paper presents a novel multi-focus image fusion algorithm that proposes to group the local connected pixels with similar colors and patterns, usually referred to as superpixels, and use them to separate the focused and de-focused regions of an image. We note that these superpixels are more expressive than individual pixels, and they carry more distinctive statistical properties when compared with other superpixels. The statistical properties of superpixels are analyzed to categorize the pixels as focused or de-focused and to estimate a focus map. A spatial consistency constraint is ensured on the initial focus map to obtain a refined map, which is used in the fusion rule to obtain a single all-in-focus image. Qualitative and quantitative evaluations are performed to assess the performance of the proposed method on a benchmark multi-focus image fusion dataset. The results show that our method produces better quality fused images than existing image fusion techniques.
Collapse
Affiliation(s)
- Areeba Ilyas
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (A.I.); (M.H.K.)
| | - Muhammad Shahid Farid
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (A.I.); (M.H.K.)
| | - Muhammad Hassan Khan
- Punjab University College of Information Technology, University of the Punjab, Lahore 54000, Pakistan; (A.I.); (M.H.K.)
| | - Marcin Grzegorzek
- Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23538 Lübeck, Germany;
| |
Collapse
|
8
|
|
9
|
Ma H, Liao Q, Zhang J, Liu S, Xue JH. An α-Matte Boundary Defocus Model-Based Cascaded Network for Multi-focus Image Fusion. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; PP:8668-8679. [PMID: 32845840 DOI: 10.1109/tip.2020.3018261] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Capturing an all-in-focus image with a single camera is difficult since the depth of field of the camera is usually limited. An alternative method to obtain the all-in-focus image is to fuse several images that are focused at different depths. However, existing multi-focus image fusion methods cannot obtain clear results for areas near the focused/defocused boundary (FDB). In this paper, a novel α-matte boundary defocus model is proposed to generate realistic training data with the defocus spread effect precisely modeled, especially for areas near the FDB. Based on this α-matte defocus model and the generated data, a cascaded boundary-aware convolutional network termed MMF-Net is proposed and trained, aiming to achieve clearer fusion results around the FDB. Specifically, the MMF-Net consists of two cascaded subnets for initial fusion and boundary fusion. These two subnets are designed to first obtain a guidance map of FDB and then refine the fusion near the FDB. Experiments demonstrate that with the help of the new α-matte boundary defocus model, the proposed MMF-Net outperforms the state-of-the-art methods both qualitatively and quantitatively.
Collapse
|