1
|
Gao Y, Cai Z, Xie X, Deng J, Dou Z, Ma X. Sparse representation for restoring images by exploiting topological structure of graph of patches. IET IMAGE PROCESSING 2025; 19. [DOI: 10.1049/ipr2.70004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/27/2024] [Accepted: 01/16/2025] [Indexed: 03/02/2025]
Abstract
AbstractImage restoration poses a significant challenge, aiming to accurately recover damaged images by delving into their inherent characteristics. Various models and algorithms have been explored by researchers to address different types of image distortions, including sparse representation, grouped sparse representation, and low‐rank self‐representation. The grouped sparse representation algorithm leverages the prior knowledge of non‐local self‐similarity and imposes sparsity constraints to maintain texture information within images. To further exploit the intrinsic properties of images, this study proposes a novel low‐rank representation‐guided grouped sparse representation image restoration algorithm. This algorithm integrates self‐representation models and trace optimization techniques to effectively preserve the original image structure, thereby enhancing image restoration performance while retaining the original texture and structural information. The proposed method was evaluated on image denoising and deblocking tasks across several datasets, demonstrating promising results.
Collapse
Affiliation(s)
- Yaxian Gao
- School of Information Engineering Shaanxi Xueqian Normal University Xi'an Shaanxi China
| | - Zhaoyuan Cai
- School of Computer Science and Technology Xidian University Xi'an Shaanxi China
| | - Xianghua Xie
- Department of Computer Science Swansea University Swansea UK
| | - Jingjing Deng
- Department of Computer Science Durham University Durham UK
| | - Zengfa Dou
- School of Information Engineering Shaanxi Xueqian Normal University Xi'an Shaanxi China
| | - Xiaoke Ma
- School of Computer Science and Technology Xidian University Xi'an Shaanxi China
| |
Collapse
|
2
|
Chen Z, He X, Zhang T, Xiong S, Ren C. Dual-stage feedback network for lightweight color image compression artifact reduction. Neural Netw 2024; 179:106555. [PMID: 39068676 DOI: 10.1016/j.neunet.2024.106555] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Revised: 05/29/2024] [Accepted: 07/18/2024] [Indexed: 07/30/2024]
Abstract
Lossy image coding techniques usually result in various undesirable compression artifacts. Recently, deep convolutional neural networks have seen encouraging advances in compression artifact reduction. However, most of them focus on the restoration of the luma channel without considering the chroma components. Besides, most deep convolutional neural networks are hard to deploy in practical applications because of their high model complexity. In this article, we propose a dual-stage feedback network (DSFN) for lightweight color image compression artifact reduction. Specifically, we propose a novel curriculum learning strategy to drive a DSFN to reduce color image compression artifacts in a luma-to-RGB manner. In the first stage, the DSFN is dedicated to reconstructing the luma channel, whose high-level features containing rich structural information are then rerouted to the second stage by a feedback connection to guide the RGB image restoration. Furthermore, we present a novel enhanced feedback block for efficient high-level feature extraction, in which an adaptive iterative self-refinement module is carefully designed to refine the low-level features progressively, and an enhanced separable convolution is advanced to exploit multiscale image information fully. Extensive experiments show the notable advantage of our DSFN over several state-of-the-art methods in both quantitative indices and visual effects with lower model complexity.
Collapse
Affiliation(s)
- Zhengxin Chen
- College of Electronic and Information Engineering, Sichuan University, Chengdu, 610065, China
| | - Xiaohai He
- College of Electronic and Information Engineering, Sichuan University, Chengdu, 610065, China
| | - Tingrong Zhang
- College of Electronic and Information Engineering, Sichuan University, Chengdu, 610065, China
| | - Shuhua Xiong
- College of Electronic and Information Engineering, Sichuan University, Chengdu, 610065, China
| | - Chao Ren
- College of Electronic and Information Engineering, Sichuan University, Chengdu, 610065, China.
| |
Collapse
|
3
|
Ma L, Zhao Y, Peng P, Tian Y. Sensitivity Decouple Learning for Image Compression Artifacts Reduction. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:3620-3633. [PMID: 38787669 DOI: 10.1109/tip.2024.3403034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2024]
Abstract
With the benefit of deep learning techniques, recent researches have made significant progress in image compression artifacts reduction. Despite their improved performances, prevailing methods only focus on learning a mapping from the compressed image to the original one but ignore the intrinsic attributes of the given compressed images, which greatly harms the performance of downstream parsing tasks. Different from these methods, we propose to decouple the intrinsic attributes into two complementary features for artifacts reduction, i.e., the compression-insensitive features to regularize the high-level semantic representations during training and the compression-sensitive features to be aware of the compression degree. To achieve this, we first employ adversarial training to regularize the compressed and original encoded features for retaining high-level semantics, and we then develop the compression quality-aware feature encoder for compression-sensitive features. Based on these dual complementary features, we propose a Dual Awareness Guidance Network (DAGN) to utilize these awareness features as transformation guidance during the decoding phase. In our proposed DAGN, we develop a cross-feature fusion module to maintain the consistency of compression-insensitive features by fusing compression-insensitive features into the artifacts reduction baseline. Our method achieves an average 2.06 dB PSNR gains on BSD500, outperforming state-of-the-art methods, and only requires 29.7 ms to process one image on BSD500. Besides, the experimental results on LIVE1 and LIU4K also demonstrate the efficiency, effectiveness, and superiority of the proposed method in terms of quantitative metrics, visual quality, and downstream machine vision tasks.
Collapse
|
4
|
Ouyang M, Chen Z. JPEG Quantized Coefficient Recovery via DCT Domain Spatial-Frequential Transformer. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:3385-3398. [PMID: 38787670 DOI: 10.1109/tip.2024.3403054] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2024]
Abstract
JPEG compression adopts the quantization of Discrete Cosine Transform (DCT) coefficients for effective bit-rate reduction, whilst the quantization could lead to a significant loss of important image details. Recovering compressed JPEG images in the frequency domain has recently garnered increasing interest, complementing the multitude of restoration techniques established in the pixel domain. However, existing DCT domain methods typically suffer from limited effectiveness in handling a wide range of compression quality factors or fall short in recovering sparse quantized coefficients and the components across different colorspaces. To address these challenges, we propose a DCT domain spatial-frequential Transformer, namely DCTransformer, for JPEG quantized coefficient recovery. Specifically, a dual-branch architecture is designed to capture both spatial and frequential correlations within the collocated DCT coefficients. Moreover, we incorporate the operation of quantization matrix embedding, which effectively allows our single model to handle a wide range of quality factors, and a luminance-chrominance alignment head that produces a unified feature map to align different-sized luminance and chrominance components. Our proposed DCTransformer outperforms the current state-of-the-art JPEG artifact removal techniques, as demonstrated by our extensive experiments.
Collapse
|
5
|
An B, Wang S, Qin F, Zhao Z, Yan R, Chen X. Adversarial Algorithm Unrolling Network for Interpretable Mechanical Anomaly Detection. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:6007-6020. [PMID: 37028350 DOI: 10.1109/tnnls.2023.3250664] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
In mechanical anomaly detection, algorithms with higher accuracy, such as those based on artificial neural networks, are frequently constructed as black boxes, resulting in opaque interpretability in architecture and low credibility in results. This article proposes an adversarial algorithm unrolling network (AAU-Net) for interpretable mechanical anomaly detection. AAU-Net is a generative adversarial network (GAN). Its generator, composed of an encoder and a decoder, is mainly produced by algorithm unrolling of a sparse coding model, which is specially designed for feature encoding and decoding of vibration signals. Thus, AAU-Net has a mechanism-driven and interpretable network architecture. In other words, it is ad hoc interpretable. Moreover, a multiscale feature visualization approach for AAU-Net is introduced to verify that meaningful features are encoded by AAU-Net, helping users to trust the detection results. The feature visualization approach enables the results of AAU-Net to be interpretable, i.e., post hoc interpretable. To verify AAU-Net's capability of feature encoding and anomaly detection, we designed and performed simulations and experiments. The results show that AAU-Net can learn signal features that match the dynamic mechanism of the mechanical system. Considering the excellent feature learning ability, unsurprisingly, AAU-Net achieves the best overall anomaly detection performance compared with other algorithms.
Collapse
|
6
|
Welker S, Chapman HN, Gerkmann T. DriftRec: Adapting Diffusion Models to Blind JPEG Restoration. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:2795-2807. [PMID: 38578859 DOI: 10.1109/tip.2024.3383776] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/07/2024]
Abstract
In this work, we utilize the high-fidelity generation abilities of diffusion models to solve blind JPEG restoration at high compression levels. We propose an elegant modification of the forward stochastic differential equation of diffusion models to adapt them to this restoration task and name our method DriftRec. Comparing DriftRec against an L2 regression baseline with the same network architecture and state-of-the-art techniques for JPEG restoration, we show that our approach can escape the tendency of other methods to generate blurry images, and recovers the distribution of clean images significantly more faithfully. For this, only a dataset of clean/corrupted image pairs and no knowledge about the corruption operation is required, enabling wider applicability to other restoration tasks. In contrast to other conditional and unconditional diffusion models, we utilize the idea that the distributions of clean and corrupted images are much closer to each other than each is to the usual Gaussian prior of the reverse process in diffusion models. Our approach therefore requires only low levels of added noise and needs comparatively few sampling steps even without further optimizations. We show that DriftRec naturally generalizes to realistic and difficult scenarios such as unaligned double JPEG compression and blind restoration of JPEGs found online, without having encountered such examples during training.
Collapse
|
7
|
Huang Z, Zhang J. Contrastive Unfolding Deraining Network. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:5155-5169. [PMID: 36112550 DOI: 10.1109/tnnls.2022.3202724] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Due to the fact that the degradation of image quality caused by rain usually affects outdoor vision tasks, image deraining becomes more and more important. Focusing on the single image deraining (SID) task, in this article, we propose a novel Contrastive Unfolding DEraining Network (CUDEN), which combines the traditional iterative algorithm and deep network, exhibiting excellent performance and nice interpretability. CUDEN transforms the challenge of locating rain streaks into discovering rain features and defines the relationship between the image and feature domains in terms of mapping pairs. To obtain the mapping pairs efficiently, we propose a dynamic multidomain translation (DMT) module for decomposing the original mapping into sub-mappings. To enhance the feature extraction capability of networks, we also propose a new serial multireceptive field fusion (SMF) block, which extracts complex and variable rain features with convolution kernels of different receptive fields. Moreover, we are the first to introduce contrastive learning to the SID task and combine it with perceptual loss to propose a new contrastive perceptual loss (CPL), which is quite generalized and greatly helpful in identifying the appropriate gradient descent direction during training. Extensive experiments on synthetic and real-world datasets demonstrate that our proposed CUDEN outperforms the state-of-the-art (SOTA) deraining networks.
Collapse
|
8
|
Cai Z, Xie X, Deng J, Dou Z, Tong B, Ma X. Image restoration with group sparse representation and low‐rank group residual learning. IET IMAGE PROCESSING 2024; 18:741-760. [DOI: 10.1049/ipr2.12982] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Accepted: 11/01/2023] [Indexed: 01/11/2025]
Abstract
AbstractImage restoration, as a fundamental research topic of image processing, is to reconstruct the original image from degraded signal using the prior knowledge of image. Group sparse representation (GSR) is powerful for image restoration; it however often leads to undesirable sparse solutions in practice. In order to improve the quality of image restoration based on GSR, the sparsity residual model expects the representation learned from degraded images to be as close as possible to the true representation. In this article, a group residual learning based on low‐rank self‐representation is proposed to automatically estimate the true group sparse representation. It makes full use of the relation among patches and explores the subgroup structures within the same group, which makes the sparse residual model have better interpretation furthermore, results in high‐quality restored images. Extensive experimental results on two typical image restoration tasks (image denoising and deblocking) demonstrate that the proposed algorithm outperforms many other popular or state‐of‐the‐art image restoration methods.
Collapse
Affiliation(s)
- Zhaoyuan Cai
- School of Computer Science and Technology Xidian University Xi'an Shaanxi China
| | - Xianghua Xie
- Department of Computer Science Swansea University Swansea UK
| | - Jingjing Deng
- Department of Computer Science Durham University Durham UK
| | - Zengfa Dou
- 20th Research Institute China Electronic Science and Technology Group Co., Ltd Xi'an Shaanxi China
| | - Bo Tong
- Xi'an Thermal Power Research Institute Co., Ltd Xi'an China
| | - Xiaoke Ma
- School of Computer Science and Technology Xidian University Xi'an Shaanxi China
| |
Collapse
|