1
|
Qiu G, Tao D, You D, Wu L. Low-illumination and noisy bridge crack image restoration by deep CNN denoiser and normalized flow module. Sci Rep 2024; 14:18270. [PMID: 39107363 PMCID: PMC11303699 DOI: 10.1038/s41598-024-69412-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2024] [Accepted: 08/05/2024] [Indexed: 08/10/2024] Open
Abstract
When applying deep learning and image processing techniques for bridge crack detection, the obtained images in real-world scenarios have severe image degradation problem. This study focuses on restoring low-illumination bridge crack images corrupted by noise to improve the accuracy of subsequent crack detection and semantic segmentation. The proposed algorithm consists of a deep CNN denoiser and a normalized flow-based brightness enhancement module. By taking the noise spectrum as an input, the deep CNN denoiser restores image at a broad range of noise levels. The normalized flow module, employs a conditional encoder and a reversible network to map the distribution of normally exposed images to a Gaussian distribution, effectively improving the image brightness. Extensive experiments have demonstrated the approach can usefully recover low-illumination images corrupted by noise compared to the state-of-the-art methods. Furthermore, the algorithm presented in this study can also be applied to other image quality restoration with high generalization and robust abilities. And the semantic segmentation accuracy of the restored image is significantly improved.
Collapse
Affiliation(s)
- Guangying Qiu
- State Key Laboratory of Rail Transit Infrastructure Performance Monitoring and Guarantee, East China Jiaotong University, Nanchang, 330013, China
| | - Dan Tao
- State Key Laboratory of Rail Transit Infrastructure Performance Monitoring and Guarantee, East China Jiaotong University, Nanchang, 330013, China.
- School of Electrical and Automation Engineering, East China Jiaotong University, Nanchang, 330013, China.
| | - Dequan You
- Fujian Communications Research Institute Co., Ltd., Fuzhou, 350000, China
| | - Linming Wu
- Fujian Communications Research Institute Co., Ltd., Fuzhou, 350000, China
| |
Collapse
|
2
|
Li Z, Zhou W, Zhou Z, Zhang S, Shi J, Shen C, Zhang J, Chi N, Dai Q. Self-supervised dynamic learning for long-term high-fidelity image transmission through unstabilized diffusive media. Nat Commun 2024; 15:1498. [PMID: 38374085 PMCID: PMC10876540 DOI: 10.1038/s41467-024-45745-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2023] [Accepted: 01/31/2024] [Indexed: 02/21/2024] Open
Abstract
Multimode fiber (MMF) which supports parallel transmission of spatially distributed information is a promising platform for remote imaging and capacity-enhanced optical communication. However, the variability of the scattering MMF channel poses a challenge for achieving long-term accurate transmission over long distances, of which static optical propagation modeling with calibrated transmission matrix or data-driven learning will inevitably degenerate. In this paper, we present a self-supervised dynamic learning approach that achieves long-term, high-fidelity transmission of arbitrary optical fields through unstabilized MMFs. Multiple networks carrying both long- and short-term memory of the propagation model variations are adaptively updated and ensembled to achieve robust image recovery. We demonstrate >99.9% accuracy in the transmission of 1024 spatial degree-of-freedom over 1 km length MMFs lasting over 1000 seconds. The long-term high-fidelity capability enables compressive encoded transfer of high-resolution video with orders of throughput enhancement, offering insights for artificial intelligence promoted diffusive spatial transmission in practical applications.
Collapse
Affiliation(s)
- Ziwei Li
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China.
- Shanghai Engineering Research Center of Low-Earth-Orbit Satellite Communication and Applications, 200433, Shanghai, China.
- Pujiang Laboratory, 200232, Shanghai, China.
| | - Wei Zhou
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China
| | - Zhanhong Zhou
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China
| | - Shuqi Zhang
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China
| | - Jianyang Shi
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China
- Shanghai Engineering Research Center of Low-Earth-Orbit Satellite Communication and Applications, 200433, Shanghai, China
| | - Chao Shen
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China
- Shanghai Engineering Research Center of Low-Earth-Orbit Satellite Communication and Applications, 200433, Shanghai, China
| | - Junwen Zhang
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China
- Shanghai Engineering Research Center of Low-Earth-Orbit Satellite Communication and Applications, 200433, Shanghai, China
| | - Nan Chi
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China.
- Shanghai Engineering Research Center of Low-Earth-Orbit Satellite Communication and Applications, 200433, Shanghai, China.
| | - Qionghai Dai
- School of Information Science and Technology, Fudan University, 200433, Shanghai, China.
- Tsinghua University, 100084, Beijing, China.
| |
Collapse
|
3
|
Image denoising in the deep learning era. Artif Intell Rev 2022. [DOI: 10.1007/s10462-022-10305-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
4
|
An Unsupervised Weight Map Generative Network for Pixel-Level Combination of Image Denoisers. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12126227] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
Image denoising is a classic but still important issue in image processing as the denoising effect has a significant impact on subsequent image processing results, such as target recognition and edge detection. In the past few decades, various denoising methods have been proposed, such as model-based and learning-based methods, and they have achieved promising results. However, no stand-alone method consistently outperforms the others in different complex imaging situations. Based on the complementary strengths of model-based and learning-based methods, in this study, we design a pixel-level image combination strategy to leverage their respective advantages for the denoised images (referred to as initial denoised images) generated by individual denoisers. The key to this combination strategy is to generate a corresponding weight map of the same size for each initial denoised image. To this end, we introduce an unsupervised weight map generative network that adjusts its parameters to generate a weight map for each initial denoised image under the guidance of our designed loss function. Using the weight maps, we are able to fully utilize the internal and external information of various denoising methods at a finer granularity, ensuring that the final combined image is close to the optimal. To the best of our knowledge, our enhancement method of combining denoised images at the pixel level is the first proposed in the image combination field. Extensive experiments demonstrate that the proposed method shows superior performance, both quantitatively and visually, and stronger generalization. Specifically, in comparison with the stand-alone denoising methods FFDNet and BM3D, our method improves the average peak signal-to-noise ratio (PSNR) by 0.18 dB to 0.83 dB on two benchmarking datasets crossing different noise levels. Its denoising effect is also greater than other competitive stand-alone methods and combination methods, and has surpassed the denoising effect of the second-best method by 0.03 dB to 1.42 dB. It should be noted that since our image combination strategy is generic, the proposed combined strategy can not only be used for image denoising but can also be extended to low-light image enhancement, image deblurring or image super-resolution.
Collapse
|
5
|
Tahir W, Wang H, Tian L. Adaptive 3D descattering with a dynamic synthesis network. LIGHT, SCIENCE & APPLICATIONS 2022; 11:42. [PMID: 35210401 PMCID: PMC8873471 DOI: 10.1038/s41377-022-00730-x] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/30/2021] [Revised: 01/22/2022] [Accepted: 02/02/2022] [Indexed: 05/11/2023]
Abstract
Deep learning has been broadly applied to imaging in scattering applications. A common framework is to train a descattering network for image recovery by removing scattering artifacts. To achieve the best results on a broad spectrum of scattering conditions, individual "expert" networks need to be trained for each condition. However, the expert's performance sharply degrades when the testing condition differs from the training. An alternative brute-force approach is to train a "generalist" network using data from diverse scattering conditions. It generally requires a larger network to encapsulate the diversity in the data and a sufficiently large training set to avoid overfitting. Here, we propose an adaptive learning framework, termed dynamic synthesis network (DSN), which dynamically adjusts the model weights and adapts to different scattering conditions. The adaptability is achieved by a novel "mixture of experts" architecture that enables dynamically synthesizing a network by blending multiple experts using a gating network. We demonstrate the DSN in holographic 3D particle imaging for a variety of scattering conditions. We show in simulation that our DSN provides generalization across a continuum of scattering conditions. In addition, we show that by training the DSN entirely on simulated data, the network can generalize to experiments and achieve robust 3D descattering. We expect the same concept can find many other applications, such as denoising and imaging in scattering media. Broadly, our dynamic synthesis framework opens up a new paradigm for designing highly adaptive deep learning and computational imaging techniques.
Collapse
Affiliation(s)
- Waleed Tahir
- Department of Electrical and Computer Engineering, Boston University, Boston, MA, 02215, USA
| | - Hao Wang
- Department of Electrical and Computer Engineering, Boston University, Boston, MA, 02215, USA
| | - Lei Tian
- Department of Electrical and Computer Engineering, Boston University, Boston, MA, 02215, USA.
- Department of Biomedical Engineering, Boston University, Boston, MA, 02215, USA.
| |
Collapse
|
6
|
Image Denoising Using a Novel Deep Generative Network with Multiple Target Images and Adaptive Termination Condition. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11114803] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Image denoising, a classic ill-posed problem, aims to recover a latent image from a noisy measurement. Over the past few decades, a considerable number of denoising methods have been studied extensively. Among these methods, supervised deep convolutional networks have garnered increasing attention, and their superior performance is attributed to their capability to learn realistic image priors from a large amount of paired noisy and clean images. However, if the image to be denoised is significantly different from the training images, it could lead to inferior results, and the networks may even produce hallucinations by using inappropriate image priors to handle an unseen noisy image. Recently, deep image prior (DIP) was proposed, and it overcame this drawback to some extent. The structure of the DIP generator network is capable of capturing the low-level statistics of a natural image using an unsupervised method with no training images other than the image itself. Compared with a supervised denoising model, the unsupervised DIP is more flexible when processing image content that must be denoised. Nevertheless, the denoising performance of DIP is usually inferior to the current supervised learning-based methods using deep convolutional networks, and it is susceptible to the over-fitting problem. To solve these problems, we propose a novel deep generative network with multiple target images and an adaptive termination condition. Specifically, we utilized mainstream denoising methods to generate two clear target images to be used with the original noisy image, enabling better guidance during the convergence process and improving the convergence speed. Moreover, we adopted the noise level estimation (NLE) technique to set a more reasonable adaptive termination condition, which can effectively solve the problem of over-fitting. Extensive experiments demonstrated that, according to the denoising results, the proposed approach significantly outperforms the original DIP method in tests on different databases. Specifically, the average peak signal-to-noise ratio (PSNR) performance of our proposed method on four databases at different noise levels is increased by 1.90 to 4.86 dB compared to the original DIP method. Moreover, our method achieves superior performance against state-of-the-art methods in terms of popular metrics, which include the structural similarity index (SSIM) and feature similarity index measurement (FSIM). Thus, the proposed method lays a good foundation for subsequent image processing tasks, such as target detection and super-resolution.
Collapse
|