1
|
Pryde MC, Rioux J, Cora AE, Volders D, Schmidt MH, Abdolell M, Bowen C, Beyea SD. Correlation of objective image quality metrics with radiologists' diagnostic confidence depends on the clinical task performed. J Med Imaging (Bellingham) 2025; 12:051803. [PMID: 40223906 PMCID: PMC11991859 DOI: 10.1117/1.jmi.12.5.051803] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2024] [Revised: 02/28/2025] [Accepted: 03/13/2025] [Indexed: 04/15/2025] Open
Abstract
Purpose Objective image quality metrics (IQMs) are widely used as outcome measures to assess acquisition and reconstruction strategies for diagnostic images. For nonpathological magnetic resonance (MR) images, these IQMs correlate to varying degrees with expert radiologists' confidence scores of overall perceived diagnostic image quality. However, it is unclear whether IQMs also correlate with task-specific diagnostic image quality or expert radiologists' confidence in performing a specific diagnostic task, which calls into question their use as surrogates for radiologist opinion. Approach 0.5 T MR images from 16 stroke patients and two healthy volunteers were retrospectively undersampled ( R = 1 to 7 × ) and reconstructed via compressed sensing. Three neuroradiologists reported the presence/absence of acute ischemic stroke (AIS) and assigned a Fazekas score describing the extent of chronic ischemic lesion burden. Neuroradiologists ranked their confidence in performing each task using a 1 to 5 Likert scale. Confidence scores were correlated with noise quality measure, the visual information fidelity criterion, the feature similarity index, root mean square error, and structural similarity (SSIM) via nonlinear regression modeling. Results Although acceleration alters image quality, neuroradiologists remain able to report pathology. All of the IQMs tested correlated to some degree with diagnostic confidence for assessing chronic ischemic lesion burden, but none correlated with diagnostic confidence in diagnosing the presence/absence of AIS due to consistent radiologist performance regardless of image degradation. Conclusions Accelerated images were helpful for understanding the ability of IQMs to assess task-specific diagnostic image quality in the context of chronic ischemic lesion burden, although not in the case of AIS diagnosis. These findings suggest that commonly used IQMs, such as the SSIM index, do not necessarily indicate an image's utility when performing certain diagnostic tasks.
Collapse
Affiliation(s)
- Michelle C. Pryde
- Dalhousie University, School of Biomedical Engineering, Halifax, Nova Scotia, Canada
| | - James Rioux
- Dalhousie University, School of Biomedical Engineering, Halifax, Nova Scotia, Canada
- Dalhousie University, Department of Diagnostic Radiology, Halifax, Nova Scotia, Canada
- Nova Scotia Health, Department of Diagnostic Imaging, Halifax, Nova Scotia, Canada
| | - Adela Elena Cora
- Dalhousie University, Department of Diagnostic Radiology, Halifax, Nova Scotia, Canada
- Nova Scotia Health, Department of Diagnostic Imaging, Halifax, Nova Scotia, Canada
| | - David Volders
- Dalhousie University, Department of Diagnostic Radiology, Halifax, Nova Scotia, Canada
- Nova Scotia Health, Department of Diagnostic Imaging, Halifax, Nova Scotia, Canada
| | - Matthias H. Schmidt
- Dalhousie University, Department of Diagnostic Radiology, Halifax, Nova Scotia, Canada
- Nova Scotia Health, Department of Diagnostic Imaging, Halifax, Nova Scotia, Canada
| | - Mohammed Abdolell
- Dalhousie University, Department of Diagnostic Radiology, Halifax, Nova Scotia, Canada
| | - Chris Bowen
- Dalhousie University, Department of Diagnostic Radiology, Halifax, Nova Scotia, Canada
- Nova Scotia Health, Department of Diagnostic Imaging, Halifax, Nova Scotia, Canada
| | - Steven D. Beyea
- Dalhousie University, School of Biomedical Engineering, Halifax, Nova Scotia, Canada
- Dalhousie University, Department of Diagnostic Radiology, Halifax, Nova Scotia, Canada
- Nova Scotia Health, Department of Diagnostic Imaging, Halifax, Nova Scotia, Canada
- IWK Health Centre, Halifax, Nova Scotia, Canada
| |
Collapse
|
2
|
Adamson PM, Desai AD, Dominic J, Varma M, Bluethgen C, Wood JP, Syed AB, Boutin RD, Stevens KJ, Vasanawala S, Pauly JM, Gunel B, Chaudhari AS. Using deep feature distances for evaluating the perceptual quality of MR image reconstructions. Magn Reson Med 2025; 94:317-330. [PMID: 39921580 PMCID: PMC12021552 DOI: 10.1002/mrm.30437] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2024] [Revised: 12/09/2024] [Accepted: 01/04/2025] [Indexed: 02/10/2025]
Abstract
PURPOSE Commonly used MR image quality (IQ) metrics have poor concordance with radiologist-perceived diagnostic IQ. Here, we develop and explore deep feature distances (DFDs)-distances computed in a lower-dimensional feature space encoded by a convolutional neural network (CNN)-as improved perceptual IQ metrics for MR image reconstruction. We further explore the impact of distribution shifts between images in the DFD CNN encoder training data and the IQ metric evaluation. METHODS We compare commonly used IQ metrics (PSNR and SSIM) to two "out-of-domain" DFDs with encoders trained on natural images, an "in-domain" DFD trained on MR images alone, and two domain-adjacent DFDs trained on large medical imaging datasets. We additionally compare these with several state-of-the-art but less commonly reported IQ metrics, visual information fidelity (VIF), noise quality metric (NQM), and the high-frequency error norm (HFEN). IQ metric performance is assessed via correlations with five expert radiologist reader scores of perceived diagnostic IQ of various accelerated MR image reconstructions. We characterize the behavior of these IQ metrics under common distortions expected during image acquisition, including their sensitivity to acquisition noise. RESULTS All DFDs and HFEN correlate more strongly with radiologist-perceived diagnostic IQ than SSIM, PSNR, and other state-of-the-art metrics, with correlations being comparable to radiologist inter-reader variability. Surprisingly, out-of-domain DFDs perform comparably to in-domain and domain-adjacent DFDs. CONCLUSION A suite of IQ metrics, including DFDs and HFEN, should be used alongside commonly-reported IQ metrics for a more holistic evaluation of MR image reconstruction perceptual quality. We also observe that general vision encoders are capable of assessing visual IQ even for MR images.
Collapse
Affiliation(s)
- Philip M. Adamson
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Arjun D. Desai
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Jeffrey Dominic
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Maya Varma
- Department of Computer Science, Stanford University, Stanford, California, USA
| | | | - Jeff P. Wood
- Austin Radiological Association, Austin, Texas, USA
| | - Ali B. Syed
- Department of Radiology, Stanford University, Stanford, California, USA
| | - Robert D. Boutin
- Department of Radiology, Stanford University, Stanford, California, USA
| | | | | | - John M. Pauly
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Beliz Gunel
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Akshay S. Chaudhari
- Department of Radiology, Stanford University, Stanford, California, USA
- Department of Biomedical Data Science, Stanford University, Stanford, California, USA
| |
Collapse
|
3
|
Komolafe TE, Zhou L, Zhao W, Wang N, Wu T. EDRAM-Net: Encoder-Decoder with Residual Attention Module Network for Low-dose Computed Tomography Reconstruction. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-6. [PMID: 40039194 DOI: 10.1109/embc53108.2024.10781702] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
The medical application of Computed Tomography (CT) is to provide detailed anatomical structures of patients without the need for invasive procedures like surgery, which is very useful for clinicians in disease diagnosis. Excessive radiation exposure can lead to the development of cancers. It is of great importance to reduce this radiation exposure by using low-dose CT (LDCT) acquisition, which is effective, but reconstructed CT images tend to be degraded, leading to the loss of vital information which is one of the most significant drawbacks of this technique. In the past few years, multiscale convolutional networks (MSCN) have been widely adopted in LDCT reconstruction to preserve vital details in reconstructed images. Based on this inspiration, we proposed an encoder-decoder network with a residual attention module (EDRAM-Net) for LDCT reconstruction. The proposed EDRAM-Net embeds the cascaded residual attention module (RAM) block into the skip connection connecting the encoder-decoder architecture. Specifically, the encoder captures and encodes details in the latent space, which is reconstructed in the decoder of the network. The RAM blocks consist of three modules: the MSCN, channel attention module (CAN), and spatial attention module (SAM). The MSCN captures features at different scales, while the CAM and SAM focus on channel and spatial details during reconstruction. The performance of EDRAM-Net evaluated on the public AAPM low-dose dataset shows that the method has improved performance in terms of estimated image quality metric compared to other comparative methods. The ablation study further revealed that using the kernel size of (7×7) for the RAM block significantly enhanced the performance of our model. It was also observed that a higher number of RAM blocks yielded improved performance but at the expense of computational complexity.
Collapse
|
4
|
Chang CW, Peng J, Safari M, Salari E, Pan S, Roper J, Qiu RLJ, Gao Y, Shu HK, Mao H, Yang X. High-resolution MRI synthesis using a data-driven framework with denoising diffusion probabilistic modeling. Phys Med Biol 2024; 69:045001. [PMID: 38241726 PMCID: PMC10839468 DOI: 10.1088/1361-6560/ad209c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2023] [Revised: 01/08/2024] [Accepted: 01/19/2024] [Indexed: 01/21/2024]
Abstract
Objective. High-resolution magnetic resonance imaging (MRI) can enhance lesion diagnosis, prognosis, and delineation. However, gradient power and hardware limitations prohibit recording thin slices or sub-1 mm resolution. Furthermore, long scan time is not clinically acceptable. Conventional high-resolution images generated using statistical or analytical methods include the limitation of capturing complex, high-dimensional image data with intricate patterns and structures. This study aims to harness cutting-edge diffusion probabilistic deep learning techniques to create a framework for generating high-resolution MRI from low-resolution counterparts, improving the uncertainty of denoising diffusion probabilistic models (DDPM).Approach. DDPM includes two processes. The forward process employs a Markov chain to systematically introduce Gaussian noise to low-resolution MRI images. In the reverse process, a U-Net model is trained to denoise the forward process images and produce high-resolution images conditioned on the features of their low-resolution counterparts. The proposed framework was demonstrated using T2-weighted MRI images from institutional prostate patients and brain patients collected in the Brain Tumor Segmentation Challenge 2020 (BraTS2020).Main results. For the prostate dataset, the bicubic interpolation model (Bicubic), conditional generative-adversarial network (CGAN), and our proposed DDPM framework improved the noise quality measure from low-resolution images by 4.4%, 5.7%, and 12.8%, respectively. Our method enhanced the signal-to-noise ratios by 11.7%, surpassing Bicubic (9.8%) and CGAN (8.1%). In the BraTS2020 dataset, the proposed framework and Bicubic enhanced peak signal-to-noise ratio from resolution-degraded images by 9.1% and 5.8%. The multi-scale structural similarity indexes were 0.970 ± 0.019, 0.968 ± 0.022, and 0.967 ± 0.023 for the proposed method, CGAN, and Bicubic, respectively.Significance. This study explores a deep learning-based diffusion probabilistic framework for improving MR image resolution. Such a framework can be used to improve clinical workflow by obtaining high-resolution images without penalty of the long scan time. Future investigation will likely focus on prospectively testing the efficacy of this framework with different clinical indications.
Collapse
Affiliation(s)
- Chih-Wei Chang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Junbo Peng
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Mojtaba Safari
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Elahheh Salari
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Shaoyan Pan
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
- Department of Biomedical Informatics, Emory University, Atlanta, GA 30308, United States of America
| | - Justin Roper
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Richard L J Qiu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Yuan Gao
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Hui-Kuo Shu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Hui Mao
- Department of Radiology and Imaging Sciences and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308, United States of America
- Department of Biomedical Informatics, Emory University, Atlanta, GA 30308, United States of America
| |
Collapse
|
5
|
Gao Q, Li Z, Zhang J, Zhang Y, Shan H. CoreDiff: Contextual Error-Modulated Generalized Diffusion Model for Low-Dose CT Denoising and Generalization. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:745-759. [PMID: 37773896 DOI: 10.1109/tmi.2023.3320812] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/01/2023]
Abstract
Low-dose computed tomography (CT) images suffer from noise and artifacts due to photon starvation and electronic noise. Recently, some works have attempted to use diffusion models to address the over-smoothness and training instability encountered by previous deep-learning-based denoising models. However, diffusion models suffer from long inference time due to a large number of sampling steps involved. Very recently, cold diffusion model generalizes classical diffusion models and has greater flexibility. Inspired by cold diffusion, this paper presents a novel COntextual eRror-modulated gEneralized Diffusion model for low-dose CT (LDCT) denoising, termed CoreDiff. First, CoreDiff utilizes LDCT images to displace the random Gaussian noise and employs a novel mean-preserving degradation operator to mimic the physical process of CT degradation, significantly reducing sampling steps thanks to the informative LDCT images as the starting point of the sampling process. Second, to alleviate the error accumulation problem caused by the imperfect restoration operator in the sampling process, we propose a novel ContextuaL Error-modulAted Restoration Network (CLEAR-Net), which can leverage contextual information to constrain the sampling process from structural distortion and modulate time step embedding features for better alignment with the input at the next time step. Third, to rapidly generalize the trained model to a new, unseen dose level with as few resources as possible, we devise a one-shot learning framework to make CoreDiff generalize faster and better using only one single LDCT image (un)paired with normal-dose CT (NDCT). Extensive experimental results on four datasets demonstrate that our CoreDiff outperforms competing methods in denoising and generalization performance, with clinically acceptable inference time. Source code is made available at https://github.com/qgao21/CoreDiff.
Collapse
|
6
|
Ohashi K, Nagatani Y, Yoshigoe M, Iwai K, Tsuchiya K, Hino A, Kida Y, Yamazaki A, Ishida T. Applicability Evaluation of Full-Reference Image Quality Assessment Methods for Computed Tomography Images. J Digit Imaging 2023; 36:2623-2634. [PMID: 37550519 PMCID: PMC10584745 DOI: 10.1007/s10278-023-00875-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2023] [Revised: 06/22/2023] [Accepted: 06/23/2023] [Indexed: 08/09/2023] Open
Abstract
Image quality assessments (IQA) are an important task for providing appropriate medical care. Full-reference IQA (FR-IQA) methods, such as peak signal-to-noise ratio (PSNR) and structural similarity (SSIM), are often used to evaluate imaging conditions, reconstruction conditions, and image processing algorithms, including noise reduction and super-resolution technology. However, these IQA methods may be inapplicable for medical images because they were designed for natural images. Therefore, this study aimed to investigate the correlation between objective assessment by some FR-IQA methods and human subjective assessment for computed tomography (CT) images. For evaluation, 210 distorted images were created from six original images using two types of degradation: noise and blur. We employed nine widely used FR-IQA methods for natural images: PSNR, SSIM, feature similarity (FSIM), information fidelity criterion (IFC), visual information fidelity (VIF), noise quality measure (NQM), visual signal-to-noise ratio (VSNR), multi-scale SSIM (MSSSIM), and information content-weighted SSIM (IWSSIM). Six observers performed subjective assessments using the double stimulus continuous quality scale (DSCQS) method. The performance of IQA methods was quantified using Pearson's linear correlation coefficient (PLCC), Spearman rank order correlation coefficient (SROCC), and root-mean-square error (RMSE). Nine FR-IQA methods developed for natural images were all strongly correlated with the subjective assessment (PLCC and SROCC > 0.8), indicating that these methods can apply to CT images. Particularly, VIF had the best values for all three items, PLCC, SROCC, and RMSE. These results suggest that VIF provides the most accurate alternative measure to subjective assessments for CT images.
Collapse
Affiliation(s)
- Kohei Ohashi
- Division of Health Sciences, Osaka University Graduate School of Medicine, Suita, Japan.
- Department of Radiology, Shiga University of Medical Science Hospital, Otsu, Japan.
| | - Yukihiro Nagatani
- Department of Radiology, Shiga University of Medical Science Hospital, Otsu, Japan
| | - Makoto Yoshigoe
- Department of Radiology, Shiga University of Medical Science Hospital, Otsu, Japan
| | - Kyohei Iwai
- Department of Radiology, Shiga University of Medical Science Hospital, Otsu, Japan
| | - Keiko Tsuchiya
- Department of Radiology, Omihachiman Community Medical Center, Omihachiman, Japan
| | - Atsunobu Hino
- Department of Radiology, Nagahama Red Cross Hospital, Nagahama, Japan
| | - Yukako Kida
- Department of Radiology, Shiga University of Medical Science Hospital, Otsu, Japan
| | - Asumi Yamazaki
- Division of Health Sciences, Osaka University Graduate School of Medicine, Suita, Japan
| | - Takayuki Ishida
- Division of Health Sciences, Osaka University Graduate School of Medicine, Suita, Japan
| |
Collapse
|
7
|
Zhang Z, Tian S, Zou W, Morin L, Zhang L. EDDMF: An Efficient Deep Discrepancy Measuring Framework for Full-Reference Light Field Image Quality Assessment. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:6426-6440. [PMID: 37966926 DOI: 10.1109/tip.2023.3329663] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2023]
Abstract
The increasing demand for immersive experience has greatly promoted the quality assessment research of Light Field Image (LFI). In this paper, we propose an efficient deep discrepancy measuring framework for full-reference light field image quality assessment. The main idea of the proposed framework is to efficiently evaluate the quality degradation of distorted LFIs by measuring the discrepancy between reference and distorted LFI patches. Firstly, a patch generation module is proposed to extract spatio-angular patches and sub-aperture patches from LFIs, which greatly reduces the computational cost. Then, we design a hierarchical discrepancy network based on convolutional neural networks to extract the hierarchical discrepancy features between reference and distorted spatio-angular patches. Besides, the local discrepancy features between reference and distorted sub-aperture patches are extracted as complementary features. After that, the angular-dominant hierarchical discrepancy features and the spatial-dominant local discrepancy features are combined to evaluate the patch quality. Finally, the quality of all patches is pooled to obtain the overall quality of distorted LFIs. To the best of our knowledge, the proposed framework is the first patch-based full-reference light field image quality assessment metric based on deep-learning technology. Experimental results on four representative LFI datasets show that our proposed framework achieves superior performance as well as lower computational complexity compared to other state-of-the-art metrics.
Collapse
|
8
|
Noordman CR, Yakar D, Bosma J, Simonis FFJ, Huisman H. Complexities of deep learning-based undersampled MR image reconstruction. Eur Radiol Exp 2023; 7:58. [PMID: 37789241 PMCID: PMC10547669 DOI: 10.1186/s41747-023-00372-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2023] [Accepted: 08/01/2023] [Indexed: 10/05/2023] Open
Abstract
Artificial intelligence has opened a new path of innovation in magnetic resonance (MR) image reconstruction of undersampled k-space acquisitions. This review offers readers an analysis of the current deep learning-based MR image reconstruction methods. The literature in this field shows exponential growth, both in volume and complexity, as the capabilities of machine learning in solving inverse problems such as image reconstruction are explored. We review the latest developments, aiming to assist researchers and radiologists who are developing new methods or seeking to provide valuable feedback. We shed light on key concepts by exploring the technical intricacies of MR image reconstruction, highlighting the importance of raw datasets and the difficulty of evaluating diagnostic value using standard metrics.Relevance statement Increasingly complex algorithms output reconstructed images that are difficult to assess for robustness and diagnostic quality, necessitating high-quality datasets and collaboration with radiologists.Key points• Deep learning-based image reconstruction algorithms are increasing both in complexity and performance.• The evaluation of reconstructed images may mistake perceived image quality for diagnostic value.• Collaboration with radiologists is crucial for advancing deep learning technology.
Collapse
Affiliation(s)
- Constant Richard Noordman
- Diagnostic Image Analysis Group, Department of Medical Imaging, Radboud University Medical Center, Nijmegen, 6525 GA, The Netherlands.
| | - Derya Yakar
- Medical Imaging Center, Departments of Radiology, Nuclear Medicine and Molecular Imaging, University Medical Center Groningen, Groningen, 9700 RB, The Netherlands
| | - Joeran Bosma
- Diagnostic Image Analysis Group, Department of Medical Imaging, Radboud University Medical Center, Nijmegen, 6525 GA, The Netherlands
| | | | - Henkjan Huisman
- Diagnostic Image Analysis Group, Department of Medical Imaging, Radboud University Medical Center, Nijmegen, 6525 GA, The Netherlands
- Department of Circulation and Medical Imaging, Norwegian University of Science and Technology, Trondheim, 7030, Norway
| |
Collapse
|
9
|
Li Z, Lan T, Li Z, Gao P. Exploring Relationships between Boltzmann Entropy of Images and Building Classification Accuracy in Land Cover Mapping. ENTROPY (BASEL, SWITZERLAND) 2023; 25:1182. [PMID: 37628212 PMCID: PMC10453494 DOI: 10.3390/e25081182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/29/2023] [Revised: 07/31/2023] [Accepted: 08/02/2023] [Indexed: 08/27/2023]
Abstract
Remote sensing images are important data sources for land cover mapping. As one of the most important artificial features in remote sensing images, buildings play a critical role in many applications, such as population estimation and urban planning. Classifying buildings quickly and accurately ensures the reliability of the above applications. It is known that the classification accuracy of buildings (usually indicated by a comprehensive index called F1) is greatly affected by image quality. However, how image quality affects building classification accuracy is still unclear. In this study, Boltzmann entropy (an index considering both compositional and configurational information, simply called BE) is employed to describe image quality, and the potential relationships between BE and F1 are explored based on images from two open-source building datasets (i.e., the WHU and Inria datasets) in three cities (i.e., Christchurch, Chicago and Austin). Experimental results show that (1) F1 fluctuates greatly in images where building proportions are small (especially in images with building proportions smaller than 1%) and (2) BE has a negative relationship with F1 (i.e., when BE becomes larger, F1 tends to become smaller). The negative relationships are confirmed using Spearman correlation coefficients (SCCs) and various confidence intervals via bootstrapping (i.e., a nonparametric statistical method). Such discoveries are helpful in deepening our understanding of how image quality affects building classification accuracy.
Collapse
Affiliation(s)
- Zhipeng Li
- Faculty of Geosciences and Environmental Engineering, Southwest Jiaotong University, Chengdu 611756, China; (Z.L.); (Z.L.)
| | - Tian Lan
- Faculty of Geosciences and Environmental Engineering, Southwest Jiaotong University, Chengdu 611756, China; (Z.L.); (Z.L.)
| | - Zhilin Li
- Faculty of Geosciences and Environmental Engineering, Southwest Jiaotong University, Chengdu 611756, China; (Z.L.); (Z.L.)
| | - Peichao Gao
- Faculty of Geographical Science, Beijing Normal University, Beijing 100875, China;
| |
Collapse
|
10
|
Luna R, Zabaleta I, Bertalmío M. State-of-the-art image and video quality assessment with a metric based on an intrinsically non-linear neural summation model. Front Neurosci 2023; 17:1222815. [PMID: 37559700 PMCID: PMC10408451 DOI: 10.3389/fnins.2023.1222815] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Accepted: 06/30/2023] [Indexed: 08/11/2023] Open
Abstract
The development of automatic methods for image and video quality assessment that correlate well with the perception of human observers is a very challenging open problem in vision science, with numerous practical applications in disciplines such as image processing and computer vision, as well as in the media industry. In the past two decades, the goal of image quality research has been to improve upon classical metrics by developing models that emulate some aspects of the visual system, and while the progress has been considerable, state-of-the-art quality assessment methods still share a number of shortcomings, like their performance dropping considerably when they are tested on a database that is quite different from the one used to train them, or their significant limitations in predicting observer scores for high framerate videos. In this work we propose a novel objective method for image and video quality assessment that is based on the recently introduced Intrinsically Non-linear Receptive Field (INRF) formulation, a neural summation model that has been shown to be better at predicting neural activity and visual perception phenomena than the classical linear receptive field. Here we start by optimizing, on a classic image quality database, the four parameters of a very simple INRF-based metric, and proceed to test this metric on three other databases, showing that its performance equals or surpasses that of the state-of-the-art methods, some of them having millions of parameters. Next, we extend to the temporal domain this INRF image quality metric, and test it on several popular video quality datasets; again, the results of our proposed INRF-based video quality metric are shown to be very competitive.
Collapse
Affiliation(s)
- Raúl Luna
- Institute of Optics, Spanish National Research Council (CSIC), Madrid, Spain
| | - Itziar Zabaleta
- Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Marcelo Bertalmío
- Institute of Optics, Spanish National Research Council (CSIC), Madrid, Spain
| |
Collapse
|
11
|
Varga D. An Optimization-Based Family of Predictive, Fusion-Based Models for Full-Reference Image Quality Assessment. J Imaging 2023; 9:116. [PMID: 37367464 DOI: 10.3390/jimaging9060116] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Revised: 06/01/2023] [Accepted: 06/05/2023] [Indexed: 06/28/2023] Open
Abstract
Given the reference (distortion-free) image, full-reference image quality assessment (FR-IQA) algorithms seek to assess the perceptual quality of the test image. Over the years, many effective, hand-crafted FR-IQA metrics have been proposed in the literature. In this work, we present a novel framework for FR-IQA that combines multiple metrics and tries to leverage the strength of each by formulating FR-IQA as an optimization problem. Following the idea of other fusion-based metrics, the perceptual quality of a test image is defined as the weighted product of several already existing, hand-crafted FR-IQA metrics. Unlike other methods, the weights are determined in an optimization-based framework and the objective function is defined to maximize the correlation and minimize the root mean square error between the predicted and ground-truth quality scores. The obtained metrics are evaluated on four popular benchmark IQA databases and compared to the state of the art. This comparison has revealed that the compiled fusion-based metrics are able to outperform other competing algorithms, including deep learning-based ones.
Collapse
|
12
|
Liu M, Huang J, Zeng D, Ding X, Paisley J. A Multiscale Approach to Deep Blind Image Quality Assessment. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:1656-1667. [PMID: 37027757 DOI: 10.1109/tip.2023.3245991] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Faithful measurement of perceptual quality is of significant importance to various multimedia applications. By fully utilizing reference images, full-reference image quality assessment (FR-IQA) methods usually achieve better prediction performance. On the other hand, no-reference image quality assessment (NR-IQA), also known as blind image quality assessment (BIQA), which does not consider the reference image, makes it a challenging but important task. Previous NR-IQA methods have focused on spatial measures at the expense of information in the available frequency bands. In this paper, we present a multiscale deep blind image quality assessment method (BIQA, M.D.) with spatial optimal-scale filtering analysis. Motivated by the multi-channel behavior of the human visual system and contrast sensitivity function, we decompose an image into a number of spatial frequency bands through multiscale filtering and extract features to map an image to its subjective quality score by applying convolutional neural network. Experimental results show that BIQA, M.D. compares well with existing NR-IQA methods and generalizes well across datasets.
Collapse
|
13
|
Bakurov I, Buzzelli M, Schettini R, Castelli M, Vanneschi L. Full-Reference Image Quality Expression via Genetic Programming. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:1458-1473. [PMID: 37027541 DOI: 10.1109/tip.2023.3244662] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Full-reference image quality measures are a fundamental tool to approximate the human visual system in various applications for digital data management: from retrieval to compression to detection of unauthorized uses. Inspired by both the effectiveness and the simplicity of hand-crafted Structural Similarity Index Measure (SSIM), in this work, we present a framework for the formulation of SSIM-like image quality measures through genetic programming. We explore different terminal sets, defined from the building blocks of structural similarity at different levels of abstraction, and we propose a two-stage genetic optimization that exploits hoist mutation to constrain the complexity of the solutions. Our optimized measures are selected through a cross-dataset validation procedure, which results in superior performance against different versions of structural similarity, measured as correlation with human mean opinion scores. We also demonstrate how, by tuning on specific datasets, it is possible to obtain solutions that are competitive with (or even outperform) more complex image quality measures.
Collapse
|
14
|
Yuan P, Bai R, Yan Y, Li S, Wang J, Cao C, Wu Q. Subjective and objective quality assessment of gastrointestinal endoscopy images: From manual operation to artificial intelligence. Front Neurosci 2023; 16:1118087. [PMID: 36865000 PMCID: PMC9971730 DOI: 10.3389/fnins.2022.1118087] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2022] [Accepted: 12/30/2022] [Indexed: 02/16/2023] Open
Abstract
Gastrointestinal endoscopy has been identified as an important tool for cancer diagnosis and therapy, particularly for treating patients with early gastric cancer (EGC). It is well known that the quality of gastroscope images is a prerequisite for achieving a high detection rate of gastrointestinal lesions. Owing to manual operation of gastroscope detection, in practice, it possibly introduces motion blur and produces low-quality gastroscope images during the imaging process. Hence, the quality assessment of gastroscope images is the key process in the detection of gastrointestinal endoscopy. In this study, we first present a novel gastroscope image motion blur (GIMB) database that includes 1,050 images generated by imposing 15 distortion levels of motion blur on 70 lossless images and the associated subjective scores produced with the manual operation of 15 viewers. Then, we design a new artificial intelligence (AI)-based gastroscope image quality evaluator (GIQE) that leverages the newly proposed semi-full combination subspace to learn multiple kinds of human visual system (HVS) inspired features for providing objective quality scores. The results of experiments conducted on the GIMB database confirm that the proposed GIQE showed more effective performance compared with its state-of-the-art peers.
Collapse
Affiliation(s)
- Peng Yuan
- The Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Endoscopy, Peking University Cancer Hospital and Institute, Beijing, China
| | - Ruxue Bai
- Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Yan Yan
- The Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Endoscopy, Peking University Cancer Hospital and Institute, Beijing, China
| | - Shijie Li
- The Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Endoscopy, Peking University Cancer Hospital and Institute, Beijing, China
| | - Jing Wang
- The Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Endoscopy, Peking University Cancer Hospital and Institute, Beijing, China
| | - Changqi Cao
- The Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Endoscopy, Peking University Cancer Hospital and Institute, Beijing, China
| | - Qi Wu
- The Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Endoscopy, Peking University Cancer Hospital and Institute, Beijing, China
| |
Collapse
|
15
|
Ullah F, Lee J, Jamil S, Kwon OJ. Subjective Assessment of Objective Image Quality Metrics Range Guaranteeing Visually Lossless Compression. SENSORS (BASEL, SWITZERLAND) 2023; 23:1297. [PMID: 36772337 PMCID: PMC9918960 DOI: 10.3390/s23031297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/19/2022] [Revised: 01/17/2023] [Accepted: 01/20/2023] [Indexed: 06/18/2023]
Abstract
The usage of media such as images and videos has been extensively increased in recent years. It has become impractical to store images and videos acquired by camera sensors in their raw form due to their huge storage size. Generally, image data is compressed with a compression algorithm and then stored or transmitted to another platform. Thus, image compression helps to reduce the storage size and transmission cost of the images and videos. However, image compression might cause visual artifacts, depending on the compression level. In this regard, performance evaluation of the compression algorithms is an essential task needed to reconstruct images with visually or near-visually lossless quality in case of lossy compression. The performance of the compression algorithms is assessed by both subjective and objective image quality assessment (IQA) methodologies. In this paper, subjective and objective IQA methods are integrated to evaluate the range of the image quality metrics (IQMs) values that guarantee the visually or near-visually lossless compression performed by the JPEG 1 standard (ISO/IEC 10918). A novel "Flicker Test Software" is developed for conducting the proposed subjective and objective evaluation study. In the flicker test, the selected test images are subjectively analyzed by subjects at different compression levels. The IQMs are calculated at the previous compression level, when the images were visually lossless for each subject. The results analysis shows that the objective IQMs with more closely packed values having the least standard deviation that guaranteed the visually lossless compression of the images with JPEG 1 are the feature similarity index measure (FSIM), the multiscale structural similarity index measure (MS-SSIM), and the information content weighted SSIM (IW-SSIM), with average values of 0.9997, 0.9970, and 0.9970 respectively.
Collapse
Affiliation(s)
| | | | | | - Oh-Jin Kwon
- Department of Electronics Engineering, Sejong University, Seoul 05006, Republic of Korea
| |
Collapse
|
16
|
Wang Z, Zhuang J, Ye S, Xu N, Xiao J, Peng C. Image Restoration Quality Assessment Based on Regional Differential Information Entropy. ENTROPY (BASEL, SWITZERLAND) 2023; 25:144. [PMID: 36673285 PMCID: PMC9857637 DOI: 10.3390/e25010144] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/29/2022] [Revised: 12/31/2022] [Accepted: 01/04/2023] [Indexed: 06/17/2023]
Abstract
With the development of image recovery models, especially those based on adversarial and perceptual losses, the detailed texture portions of images are being recovered more naturally. However, these restored images are similar but not identical in detail texture to their reference images. With traditional image quality assessment methods, results with better subjective perceived quality often score lower in objective scoring. Assessment methods suffer from subjective and objective inconsistencies. This paper proposes a regional differential information entropy (RDIE) method for image quality assessment to address this problem. This approach allows better assessment of similar but not identical textural details and achieves good agreement with perceived quality. Neural networks are used to reshape the process of calculating information entropy, improving the speed and efficiency of the operation. Experiments conducted with this study's image quality assessment dataset and the PIPAL dataset show that the proposed RDIE method yields a high degree of agreement with people's average opinion scores compared with other image quality assessment metrics, proving that RDIE can better quantify the perceived quality of images.
Collapse
Affiliation(s)
- Zhiyu Wang
- Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315211, China
| | - Jiayan Zhuang
- Ningbo Institute of Industrial Technology, Chinese Academy of Sciences, Ningbo 315201, China
| | - Sichao Ye
- Ningbo Institute of Industrial Technology, Chinese Academy of Sciences, Ningbo 315201, China
| | - Ningyuan Xu
- College of Materials Science and Opto-Electronic Technology, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Jiangjian Xiao
- Ningbo Institute of Industrial Technology, Chinese Academy of Sciences, Ningbo 315201, China
| | - Chengbin Peng
- Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315211, China
| |
Collapse
|
17
|
Zhang X, Li Z, Nan N, Wang X. Super-resolution reconstruction algorithm for optical-resolution photoacoustic microscopy images based on sparsity and deconvolution. OPTICS EXPRESS 2023; 31:598-609. [PMID: 36606995 DOI: 10.1364/oe.471807] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/27/2022] [Accepted: 11/08/2022] [Indexed: 06/17/2023]
Abstract
The lateral resolution of the optical-resolution photoacoustic microscopy (OR-PAM) system depends on the focusing diameter of the probe beam. By increasing the numerical aperture (NA) of optical focusing, the lateral resolution of OR-PAM can be improved. However, the increase in NA results in smaller working distances, and the entire imaging system becomes very sensitive to small optical imperfections. The existing deconvolution-based algorithms are limited by the image signal-to-noise ratio when improving the resolution of OR-PAM images. In this paper, a super-resolution reconstruction algorithm for OR-PAM images based on sparsity and deconvolution is proposed. The OR-PAM image is sparsely reconstructed according to the constructed loss function, which utilizes the sparsity of the image to combat the decrease in the resolution. The gradient accelerated Landweber iterative algorithm is used to deconvolve to obtain high-resolution OR-PAM images. Experimental results show that the proposed algorithm can improve the resolution of mouse retinal images by approximately 1.7 times without increasing the NA of the imaging system. In addition, compared to the Richardson-Lucy algorithm, the proposed algorithm can further improve the image resolution and maintain better imaging quality, which provides a foundation for the development of OR-PAM in clinical research.
Collapse
|
18
|
Lee W, Cho E, Kim W, Choi H, Beck KS, Yoon HJ, Baek J, Choi JH. No-reference perceptual CT image quality assessment based on a self-supervised learning framework. MACHINE LEARNING: SCIENCE AND TECHNOLOGY 2022. [DOI: 10.1088/2632-2153/aca87d] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022] Open
Abstract
Abstract
Accurate image quality assessment (IQA) is crucial to optimize computed tomography (CT) image protocols while keeping the radiation dose as low as reasonably achievable. In the medical domain, IQA is based on how well an image provides a useful and efficient presentation necessary for physicians to make a diagnosis. Moreover, IQA results should be consistent with radiologists’ opinions on image quality, which is accepted as the gold standard for medical IQA. As such, the goals of medical IQA are greatly different from those of natural IQA. In addition, the lack of pristine reference images or radiologists’ opinions in a real-time clinical environment makes IQA challenging. Thus, no-reference IQA (NR-IQA) is more desirable in clinical settings than full-reference IQA (FR-IQA). Leveraging an innovative self-supervised training strategy for object detection models by detecting virtually inserted objects with geometrically simple forms, we propose a novel NR-IQA method, named deep detector IQA (D2IQA), that can automatically calculate the quantitative quality of CT images. Extensive experimental evaluations on clinical and anthropomorphic phantom CT images demonstrate that our D2IQA is capable of robustly computing perceptual image quality as it varies according to relative dose levels. Moreover, when considering the correlation between the evaluation results of IQA metrics and radiologists’ quality scores, our D2IQA is marginally superior to other NR-IQA metrics and even shows performance competitive with FR-IQA metrics.
Collapse
|
19
|
Duan H, Min X, Zhu Y, Zhai G, Yang X, Le Callet P. Confusing Image Quality Assessment: Toward Better Augmented Reality Experience. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:7206-7221. [PMID: 36367913 DOI: 10.1109/tip.2022.3220404] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
With the development of multimedia technology, Augmented Reality (AR) has become a promising next-generation mobile platform. The primary value of AR is to promote the fusion of digital contents and real-world environments, however, studies on how this fusion will influence the Quality of Experience (QoE) of these two components are lacking. To achieve better QoE of AR, whose two layers are influenced by each other, it is important to evaluate its perceptual quality first. In this paper, we consider AR technology as the superimposition of virtual scenes and real scenes, and introduce visual confusion as its basic theory. A more general problem is first proposed, which is evaluating the perceptual quality of superimposed images, i.e., confusing image quality assessment. A ConFusing Image Quality Assessment (CFIQA) database is established, which includes 600 reference images and 300 distorted images generated by mixing reference images in pairs. Then a subjective quality perception experiment is conducted towards attaining a better understanding of how humans perceive the confusing images. Based on the CFIQA database, several benchmark models and a specifically designed CFIQA model are proposed for solving this problem. Experimental results show that the proposed CFIQA model achieves state-of-the-art performance compared to other benchmark models. Moreover, an extended ARIQA study is further conducted based on the CFIQA study. We establish an ARIQA database to better simulate the real AR application scenarios, which contains 20 AR reference images, 20 background (BG) reference images, and 560 distorted images generated from AR and BG references, as well as the correspondingly collected subjective quality ratings. Three types of full-reference (FR) IQA benchmark variants are designed to study whether we should consider the visual confusion when designing corresponding IQA algorithms. An ARIQA metric is finally proposed for better evaluating the perceptual quality of AR images. Experimental results demonstrate the good generalization ability of the CFIQA model and the state-of-the-art performance of the ARIQA model. The databases, benchmark models, and proposed metrics are available at: https://github.com/DuanHuiyu/ARIQA.
Collapse
|
20
|
Bowen EFW, Rodriguez AM, Sowinski DR, Granger R. Visual stream connectivity predicts assessments of image quality. J Vis 2022; 22:4. [PMID: 36219145 PMCID: PMC9580224 DOI: 10.1167/jov.22.11.4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022] Open
Abstract
Despite extensive study of early vision, new and unexpected mechanisms continue to be identified. We introduce a novel formal treatment of the psychophysics of image similarity, derived directly from straightforward connectivity patterns in early visual pathways. The resulting differential geometry formulation is shown to provide accurate and explanatory accounts of human perceptual similarity judgments. The direct formal predictions are then shown to be further improved via simple regression on human behavioral reports, which in turn are used to construct more elaborate hypothesized neural connectivity patterns. It is shown that the predictive approaches introduced here outperform a standard successful published measure of perceived image fidelity; moreover, the approach provides clear explanatory principles of these similarity findings.
Collapse
Affiliation(s)
- Elijah F W Bowen
- Brain Engineering Laboratory, Department of Psychological and Brain Sciences, Dartmouth, Hanover, NH, USA.,
| | - Antonio M Rodriguez
- Brain Engineering Laboratory, Department of Psychological and Brain Sciences, Dartmouth, Hanover, NH, USA.,
| | - Damian R Sowinski
- Brain Engineering Laboratory, Department of Psychological and Brain Sciences, Dartmouth, Hanover, NH, USA.,
| | - Richard Granger
- Brain Engineering Laboratory, Department of Psychological and Brain Sciences, Dartmouth, Hanover, NH, USA.,
| |
Collapse
|
21
|
Full-Reference Image Quality Assessment Based on an Optimal Linear Combination of Quality Measures Selected by Simulated Annealing. J Imaging 2022; 8:jimaging8080224. [PMID: 36005467 PMCID: PMC9409967 DOI: 10.3390/jimaging8080224] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2022] [Revised: 08/17/2022] [Accepted: 08/18/2022] [Indexed: 11/19/2022] Open
Abstract
Digital images can be distorted or contaminated by noise in various steps of image acquisition, transmission, and storage. Thus, the research of such algorithms, which can evaluate the perceptual quality of digital images consistent with human quality judgement, is a hot topic in the literature. In this study, an image quality assessment (IQA) method is introduced that predicts the perceptual quality of a digital image by optimally combining several IQA metrics. To be more specific, an optimization problem is defined first using the weighted sum of a few IQA metrics. Subsequently, the optimal values of the weights are determined by minimizing the root mean square error between the predicted and ground-truth scores using the simulated annealing algorithm. The resulted optimization-based IQA metrics were assessed and compared to other state-of-the-art methods on four large, widely applied benchmark IQA databases. The numerical results empirically corroborate that the proposed approach is able to surpass other competing IQA methods.
Collapse
|
22
|
Ling Y, Zhou F, Guo K, Xue JH. ASSP: An adaptive sample statistics-based pooling for full-reference image quality assessment. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.12.098] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
23
|
|
24
|
Ruan H, Tan Z, Chen L, Wan W, Cao J. Efficient sub-pixel convolutional neural network for terahertz image super-resolution. OPTICS LETTERS 2022; 47:3115-3118. [PMID: 35709064 DOI: 10.1364/ol.454267] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Accepted: 05/23/2022] [Indexed: 06/15/2023]
Abstract
Terahertz waves are electromagnetic waves located at 0.1-10 THz, and terahertz imaging technology can be applied to security inspection, biomedicine, non-destructive testing of materials, and other fields. At present, terahertz images have unclear data and rough edges. Therefore, improving the resolution of terahertz images is one of the current hot research topics. This paper proposes an efficient terahertz image super-resolution model, which is used to extract low-resolution (LR) image features and learn the mapping of LR images to high-resolution (HR) images, and then introduce an attention mechanism to let the network pay attention to more information features. Finally, we use sub-pixel convolution to learn a set of scaling filters to upgrade the final LR feature map to an HR output, which not only reduces the model complexity, but also improves the quality of the terahertz image. The resolution reaches 31.67 db on the peak signal-to-noise ratio (PSNR) index and 0.86 on the structural similarity (SSIM) index. Experiments show that the efficient sub-pixel convolutional neural network used in this article achieves better accuracy and visual improvement compared with other terahertz image super-resolution algorithms.
Collapse
|
25
|
An object-based sparse representation model for spatiotemporal image fusion. Sci Rep 2022; 12:5021. [PMID: 35322054 PMCID: PMC8943014 DOI: 10.1038/s41598-022-08728-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2021] [Accepted: 03/10/2022] [Indexed: 11/12/2022] Open
Abstract
Many algorithms have been proposed for spatiotemporal image fusion on simulated data, yet only a few deal with spectral changes in real satellite images. An innovative spatiotemporal sparse representation (STSR) image fusion approach is introduced in this study to generate global dense high spatial and temporal resolution images from real satellite images. It aimed to minimize the data gap, especially when fine spatial resolution images are unavailable for a specific period. The proposed approach uses a set of real coarse- and fine-spatial resolution satellite images acquired simultaneously and another coarse image acquired at a different time to predict the corresponding unknown fine image. During the fusion process, pixels located between object classes with different spectral responses are more vulnerable to spectral distortion. Therefore, firstly, a rule-based fuzzy classification algorithm is used in STSR to classify input data and extract accurate edge candidates. Then, an object-based estimation of physical constraints and brightness shift between input data is utilized to construct the proposed sparse representation (SR) model that can deal with real input satellite images. Initial rules to adjust spatial covariance and equalize spectral response of object classes between input images are introduced as prior information to the model, followed by an optimization step to improve the STSR approach. The proposed method is applied to real fine Sentinel-2 and coarse Landsat-8 satellite data. The results showed that introducing objects in the fusion process improved spatial detail, especially over the edge candidates, and eliminated spectral distortion by preserving the spectral continuity of extracted objects. Experiments revealed the promising performance of the proposed object-based STSR image fusion approach based on its quantitative results, where it preserved almost 96.9% and 93.8% of the spectral detail over the smooth and urban areas, respectively.
Collapse
|
26
|
A Crypto-Steganography Approach for Hiding Ransomware within HEVC Streams in Android IoT Devices. SENSORS 2022; 22:s22062281. [PMID: 35336452 PMCID: PMC8955722 DOI: 10.3390/s22062281] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Revised: 03/07/2022] [Accepted: 03/14/2022] [Indexed: 11/25/2022]
Abstract
Steganography is a vital security approach that hides any secret content within ordinary data, such as multimedia. This hiding aims to achieve the confidentiality of the IoT secret data; whether it is benign or malicious (e.g., ransomware) and for defensive or offensive purposes. This paper introduces a hybrid crypto-steganography approach for ransomware hiding within high-resolution video frames. This proposed approach is based on hybridizing an AES (advanced encryption standard) algorithm and LSB (least significant bit) steganography process. Initially, AES encrypts the secret Android ransomware data, and then LSB embeds it based on random selection criteria for the cover video pixels. This research examined broad objective and subjective quality assessment metrics to evaluate the performance of the proposed hybrid approach. We used different sizes of ransomware samples and different resolutions of HEVC (high-efficiency video coding) frames to conduct simulation experiments and comparison studies. The assessment results prove the superior efficiency of the introduced hybrid crypto-steganography approach compared to other existing steganography approaches in terms of (a) achieving the integrity of the secret ransomware data, (b) ensuring higher imperceptibility of stego video frames, (3) introducing a multi-level security approach using the AES encryption in addition to the LSB steganography, (4) performing randomness embedding based on RPS (random pixel selection) for concealing secret ransomware bits, (5) succeeding in fully extracting the ransomware data at the receiver side, (6) obtaining strong subjective and objective qualities for all tested evaluation metrics, (7) embedding different sizes of secret data at the same time within the video frame, and finally (8) passing the security scanning tests of 70 antivirus engines without detecting the existence of the embedded ransomware.
Collapse
|
27
|
A skewness reformed complex diffusion based unsharp masking for the restoration and enhancement of Poisson noise corrupted mammograms. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2021.103421] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
28
|
Glossiness Index of Objects in Halftone Color Images Based on Structure and Appearance Distortion. J Imaging 2022; 8:jimaging8030059. [PMID: 35324614 PMCID: PMC8954649 DOI: 10.3390/jimaging8030059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2021] [Revised: 02/21/2022] [Accepted: 02/22/2022] [Indexed: 02/01/2023] Open
Abstract
This paper proposes an objective glossiness index for objects in halftone color images. In the proposed index, we consider the characteristics of the human visual system (HVS) and associate the image’s structure distortion and statistical information. According to the difference in the number of strategies adopted by the HVS in judging the difference between images, it is divided into single and multi-strategy modeling. In this study, we advocate multiple strategies to determine glossy or non-glossy quality. We assumed that HVS used different visual mechanisms to evaluate glossy and non-glossy objects. For non-glossy images, the image structure dominated, so the HVS tried to use structural information to judge distortion (a strategy based on structural distortion detection). For glossy images, the glossy appearance dominated; thus, the HVS tried to search for the glossiness difference (an appearance-based strategy). Herein, we present an index for glossiness assessment that attempts to explicitly model the structural dissimilarity and appearance distortion. We used the contrast sensitivity function to account for the mechanism of halftone images when viewed by the human eye. We estimated the structure distortion for the first strategy by using local luminance and contrast masking; meanwhile, local statistics changing in the spatial frequency components for skewness and standard deviation were used to estimate the appearance distortion for the second strategy. Experimental results showed that these two mixed-distortion measurement strategies performed well in consistency with the subjective ratings of glossiness in halftone color images.
Collapse
|
29
|
Meng C, An P, Huang X, Yang C, Chen Y. Image Quality Evaluation of Light Field Image Based on Macro-Pixels and Focus Stack. Front Comput Neurosci 2022; 15:768021. [PMID: 35126077 PMCID: PMC8810542 DOI: 10.3389/fncom.2021.768021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2021] [Accepted: 12/15/2021] [Indexed: 11/13/2022] Open
Abstract
Due to the complex angular-spatial structure, light field (LF) image processing faces more opportunities and challenges than ordinary image processing. The angular-spatial structure loss of LF images can be reflected from their various representations. The angular and spatial information penetrate each other, so it is necessary to extract appropriate features to analyze the angular-spatial structure loss of distorted LF images. In this paper, a LF image quality evaluation model, namely MPFS, is proposed based on the prediction of global angular-spatial distortion of macro-pixels and the evaluation of local angular-spatial quality of the focus stack. Specifically, the angular distortion of the LF image is first evaluated through the luminance and chrominance of macro-pixels. Then, we use the saliency of spatial texture structure to pool an array of predicted values of angular distortion to obtain the predicted value of global distortion. Secondly, the local angular-spatial quality of the LF image is analyzed through the principal components of the focus stack. The focalizing structure damage caused by the angular-spatial distortion is calculated using the features of corner and texture structures. Finally, the global and local angular-spatial quality evaluation models are combined to realize the evaluation of the overall quality of the LF image. Extensive comparative experiments show that the proposed method has high efficiency and precision.
Collapse
|
30
|
Lei F, Li S, Xie S, Liu J. Subjective and Objective Quality Assessment of Swimming Pool Images. Front Neurosci 2022; 15:766762. [PMID: 35087371 PMCID: PMC8787121 DOI: 10.3389/fnins.2021.766762] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2021] [Accepted: 11/08/2021] [Indexed: 11/13/2022] Open
Abstract
As the research basis of image processing and computer vision research, image quality evaluation (IQA) has been widely used in different visual task fields. As far as we know, limited efforts have been made to date to gather swimming pool image databases and benchmark reliable objective quality models, so far. To filled this gap, in this paper we reported a new database of underwater swimming pool images for the first time, which is composed of 1500 images and associated subjective ratings recorded by 16 inexperienced observers. In addition, we proposed a main target area extraction and multi-feature fusion image quality assessment (MM-IQA) for a swimming pool environment, which performs pixel-level fusion for multiple features of the image on the premise of highlighting important detection objects. Meanwhile, a variety of well-established full-reference (FR) quality evaluation methods and partial no-reference (NR) quality evaluation algorithms are selected to verify the database we created. Extensive experimental results show that the proposed algorithm is superior to the most advanced image quality models in performance evaluation and the outcomes of subjective and objective quality assessment of most methods involved in the comparison have good correlation and consistency, which further indicating indicates that the establishment of a large-scale pool image quality assessment database is of wide applicability and importance.
Collapse
|
31
|
Zhang H, Hu X, Gou R, Zhang L, Zheng B, Shen Z. Rich Structural Index for Stereoscopic Image Quality Assessment. SENSORS 2022; 22:s22020499. [PMID: 35062460 PMCID: PMC8780543 DOI: 10.3390/s22020499] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/29/2021] [Revised: 01/05/2022] [Accepted: 01/07/2022] [Indexed: 02/04/2023]
Abstract
The human visual system (HVS), affected by viewing distance when perceiving the stereo image information, is of great significance to study of stereoscopic image quality assessment. Many methods of stereoscopic image quality assessment do not have comprehensive consideration for human visual perception characteristics. In accordance with this, we propose a Rich Structural Index (RSI) for Stereoscopic Image objective Quality Assessment (SIQA) method based on multi-scale perception characteristics. To begin with, we put the stereo pair into the image pyramid based on Contrast Sensitivity Function (CSF) to obtain sensitive images of different resolution. Then, we obtain local Luminance and Structural Index (LSI) in a locally adaptive manner on gradient maps which consider the luminance masking and contrast masking. At the same time we use Singular Value Decomposition (SVD) to obtain the Sharpness and Intrinsic Structural Index (SISI) to effectively capture the changes introduced in the image (due to distortion). Meanwhile, considering the disparity edge structures, we use gradient cross-mapping algorithm to obtain Depth Texture Structural Index (DTSI). After that, we apply the standard deviation method for the above results to obtain contrast index of reference and distortion components. Finally, for the loss caused by the randomness of the parameters, we use Support Vector Machine Regression based on Genetic Algorithm (GA-SVR) training to obtain the final quality score. We conducted a comprehensive evaluation with state-of-the-art methods on four open databases. The experimental results show that the proposed method has stable performance and strong competitive advantage.
Collapse
Affiliation(s)
- Hua Zhang
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China; (H.Z.); (X.H.); (R.G.); (B.Z.); (Z.S.)
- Key Laboratory of Network Multimedia Technology of Zhejiang Province, Zhejiang University, Hangzhou 310018, China
| | - Xinwen Hu
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China; (H.Z.); (X.H.); (R.G.); (B.Z.); (Z.S.)
| | - Ruoyun Gou
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China; (H.Z.); (X.H.); (R.G.); (B.Z.); (Z.S.)
| | - Lingjun Zhang
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China; (H.Z.); (X.H.); (R.G.); (B.Z.); (Z.S.)
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou Dianzi University, Hangzhou 310018, China
- Correspondence:
| | - Bolun Zheng
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China; (H.Z.); (X.H.); (R.G.); (B.Z.); (Z.S.)
| | - Zhuonan Shen
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China; (H.Z.); (X.H.); (R.G.); (B.Z.); (Z.S.)
| |
Collapse
|
32
|
Rodriguez-Gallo Y, Orozco-Morales R, Marlen Perez-Diaz. Analysis of objective quality metrics in computed tomography images affected by metal artifacts. BIOMED ENG-BIOMED TE 2021; 67:1-9. [PMID: 34964320 DOI: 10.1515/bmt-2020-0244] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2020] [Accepted: 11/26/2021] [Indexed: 11/15/2022]
Abstract
Image quality (IQ) assessment plays an important role in the medical world. New methods to evaluate image quality have been developed, but their application in the context of computer tomography is yet limited. In this paper the performance of fifteen well-known full reference (FR) IQ metrics is compared with human judgment using images affected by metal artifacts and processed with metal artifact reduction methods from a phantom. Five region of interest with different sizes were selected. IQ was evaluated by seven experienced radiologists completely blinded to the information. To measure the correlation between FR-IQ, and the score assigned by radiologists non-parametric Spearman rank-order correlation coefficient and Kendall's Rank-order Correlation coefficient were used; so as root mean square error and the mean absolute error to measure the prediction accuracy. Cohen's kappa was employed with the purpose of assessing inter-observer agreement. The metrics GMSD, IWMSE, IWPSNR, WSNR and OSS-PSNR were the best ranked. Inter-observer agreement was between 0.596 and 0.954, with p<0.001 in all study. The objective scores predicted by these methods correlate consistently with the subjective evaluations. The application of this metrics will make possible a better evaluation of metal artifact reduction algorithms in future works.
Collapse
Affiliation(s)
- Yakdiel Rodriguez-Gallo
- Departamento de Control Automático, Universidad Central 'Marta Abreu' de Las Villas, Santa Clara, Cuba
| | - Ruben Orozco-Morales
- Departamento de Control Automático, Universidad Central 'Marta Abreu' de Las Villas, Santa Clara, Cuba
| | - Marlen Perez-Diaz
- Departamento de Control Automático, Universidad Central 'Marta Abreu' de Las Villas, Santa Clara, Cuba
| |
Collapse
|
33
|
Merzougui N, Djerou L. Multi-gene Genetic Programming based Predictive Models for Full-reference Image Quality Assessment. J Imaging Sci Technol 2021. [DOI: 10.2352/j.imagingsci.technol.2021.65.6.060409] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/01/2022]
Abstract
Abstract Many objective quality metrics for assessing the visual quality of images have been developed during the last decade. A simple way to fine tune the efficiency of assessment is through permutation and combination of these metrics. The goal of this fusion approach
is to take advantage of the metrics utilized and minimize the influence of their drawbacks. In this paper, a symbolic regression technique using an evolutionary algorithm known as multi-gene genetic programming (MGGP) is applied for predicting subject scores of images in datasets using a combination
of objective scores of a set of image quality metrics (IQM). By learning from image datasets, the MGGP algorithm can determine appropriate image quality metrics, from 21 metrics utilized, whose objective scores employed as predictors in the symbolic regression model, by optimizing simultaneously
two competing objectives of model ‘goodness of fit’ to data and model ‘complexity’. Six large image databases (namely LIVE, CSIQ, TID2008, TID2013, IVC and MDID) that are available in public domain are used for learning and testing the predictive models, according the
k-fold-cross-validation and the cross dataset strategies. The proposed approach is compared against state-of-the-art objective image quality assessment approaches. Results of comparison reveal that the proposed approach outperforms other state-of-the-art recently developed fusion approaches.
Collapse
Affiliation(s)
| | - Leila Djerou
- LESIA laboratory, University Mohamed Khider Biskra, Algeria
| |
Collapse
|
34
|
Simi V, Reddy Edla D, Joseph J. A no-reference metric to assess quality of denoising for Magnetic Resonance images. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2021.102962] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
35
|
Kumar A, Srivastava S. Restoration and enhancement of breast ultrasound images using extended complex diffusion based unsharp masking. Proc Inst Mech Eng H 2021; 236:12-29. [PMID: 34405743 DOI: 10.1177/09544119211039317] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Ultrasound is a well-known imaging modality for the interpretation of breast cancer. It is playing very important role for breast cancer detection that are missed by mammograms. The image acquisition is usually affected by the presence of noise, artifacts, and distortion. To overcome such type of issues, there is a need of image restoration and enhancement to improve the quality of image. This paper proposes a single framework for denoising and enhancement of ultrasound images, where a smoothing filter is replaced with an extended complex diffusion-based filter in an unsharp masking technique. The performance evaluation of the proposed method is tested on real ultrasound breast cancer images database and synthetic ultrasound image. The performance evaluation comprises qualitative and quantitative evaluation along with comparative analysis of pre-existing and proposed method. The quantitative evaluation metrics are mean squared error, peak-signal-to-noise ratio, correlation parameter, normalized absolute error, universal quality index, similarity structure index, edge preservation index, a measure of enhancement, a measure of enhancement by entropy, and second derivative like measurement. The result specifies that the proposed method is better suited approach for the removal of speckle noise which follows Rayleigh distribution, restoration of information, enhancement of abnormalities, and proper edge preservation.
Collapse
Affiliation(s)
- Abhinav Kumar
- Electronics and Communication Department, National Institute of Technology Patna, Patna, Bihar, India
| | - Subodh Srivastava
- Electronics and Communication Department, National Institute of Technology Patna, Patna, Bihar, India
| |
Collapse
|
36
|
RAVA: Region-Based Average Video Quality Assessment. SENSORS 2021; 21:s21165489. [PMID: 34450931 PMCID: PMC8401697 DOI: 10.3390/s21165489] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/28/2021] [Revised: 08/06/2021] [Accepted: 08/11/2021] [Indexed: 12/04/2022]
Abstract
Video has become the most popular medium of communication over the past decade, with nearly 90 percent of the bandwidth on the Internet being used for video transmission. Thus, evaluating the quality of an acquired or compressed video has become increasingly important. The goal of video quality assessment (VQA) is to measure the quality of a video clip as perceived by a human observer. Since manually rating every video clip to evaluate quality is infeasible, researchers have attempted to develop various quantitative metrics that estimate the perceptual quality of video. In this paper, we propose a new region-based average video quality assessment (RAVA) technique extending image quality assessment (IQA) metrics. In our experiments, we extend two full-reference (FR) image quality metrics to measure the feasibility of the proposed RAVA technique. Results on three different datasets show that our RAVA method is practical in predicting objective video scores.
Collapse
|
37
|
Chen L, Han J, Tian F. Colorization of fusion image of infrared and visible images based on parallel generative adversarial network approach. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2021. [DOI: 10.3233/jifs-210987] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Fusing the infrared (IR) and visible images has many advantages and can be applied to applications such as target detection and recognition. Colors can give more accurate and distinct features, but the low resolution and low contrast of fused images make this a challenge task. In this paper, we proposed a method based on parallel generative adversarial networks (GANs) to address the challenge. We used IR image, visible image and fusion image as ground truth of ‘L’, ‘a’ and ‘b’ of the Lab model. Through the parallel GANs, we can gain the Lab data which can be converted to RGB image. We adopt TNO and RoadScene data sets to verify our method, and compare with five objective evaluation parameters obtained by other three methods based on deep learning (DL). It is demonstrated that the proposed approach is able to achieve better performance against state-of-arts methods.
Collapse
Affiliation(s)
- Lei Chen
- Xi’an Technological University, Xi’an, Shaanxi Province, China
| | - Jun Han
- Xi’an Technological University, Xi’an, Shaanxi Province, China
| | - Feng Tian
- Bournemouth University, Poole, BH12 5BB, UK
| |
Collapse
|
38
|
Lévêque L, Outtas M, Liu H, Zhang L. Comparative study of the methodologies used for subjective medical image quality assessment. Phys Med Biol 2021; 66. [PMID: 34225264 DOI: 10.1088/1361-6560/ac1157] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2021] [Accepted: 07/05/2021] [Indexed: 11/12/2022]
Abstract
Healthcare professionals have been increasingly viewing medical images and videos in their routine clinical practice, and this in a wide variety of environments. Both the perception and interpretation of medical visual information, across all branches of practice or medical specialties (e.g. diagnostic, therapeutic, or surgical medicine), career stages, and practice settings (e.g. emergency care), appear to be critical for patient care. However, medical images and videos are not self-explanatory and, therefore, need to be interpreted by humans, i.e. medical experts. In addition, various types of degradations and artifacts may appear during image acquisition or processing, and consequently affect medical imaging data. Such distortions tend to impact viewers' quality of experience, as well as their clinical practice. It is accordingly essential to better understand how medical experts perceive the quality of visual content. Thankfully, progress has been made in the recent literature towards such understanding. In this article, we present an up-to-date state-of the-art of relatively recent (i.e. not older than ten years old) existing studies on the subjective quality assessment of medical images and videos, as well as research works using task-based approaches. Furthermore, we discuss the merits and drawbacks of the methodologies used, and we provide recommendations about experimental designs and statistical processes to evaluate the perception of medical images and videos for future studies, which could then be used to optimise the visual experience of image readers in real clinical practice. Finally, we tackle the issue of the lack of available annotated medical image and video quality databases, which appear to be indispensable for the development of new dedicated objective metrics.
Collapse
Affiliation(s)
- Lucie Lévêque
- Nantes Laboratory of Digital Sciences (LS2N), University of Nantes, Nantes, France
| | - Meriem Outtas
- Department of Industrial Computer Science and Electronics, National Institute of Applied Sciences, Rennes, France
| | - Hantao Liu
- School of Computer Science and Informatics, Cardiff University, Cardiff, United Kingdom
| | - Lu Zhang
- Department of Industrial Computer Science and Electronics, National Institute of Applied Sciences, Rennes, France
| |
Collapse
|
39
|
Ma X, Xu H, Restrepo CM, Arce GR. Multi-objective optimization for structured illumination in dynamic x-ray tomosynthesis. APPLIED OPTICS 2021; 60:6177-6188. [PMID: 34613284 DOI: 10.1364/ao.428871] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/26/2021] [Accepted: 06/22/2021] [Indexed: 06/13/2023]
Abstract
Dynamic coded x-ray tomosynthesis (CXT) uses a set of encoded x-ray sources to interrogate objects lying on a moving conveyor mechanism. The object is reconstructed from the encoded measurements received by the uniform linear array detectors. We propose a multi-objective optimization (MO) method for structured illuminations to balance the reconstruction quality and radiation dose in a dynamic CXT system. The MO framework is established based on a dynamic sensing geometry with binary coding masks. The Strength Pareto Evolutionary Algorithm 2 is used to solve the MO problem by jointly optimizing the coding masks, locations of x-ray sources, and exposure moments. Computational experiments are implemented to assess the proposed MO method. They show that the proposed strategy can obtain a set of Pareto optimal solutions with different levels of radiation dose and better reconstruction quality than the initial setting.
Collapse
|
40
|
Image Quality Assessment without Reference by Combining Deep Learning-Based Features and Viewing Distance. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11104661] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
An abundance of objective image quality metrics have been introduced in the literature. One important essential aspect that perceived image quality is dependent on is the viewing distance from the observer to the image. We introduce in this study a novel image quality metric able to estimate the quality of a given image without reference for different viewing distances between the image and the observer. We first select relevant patches from the image using saliency information. For each patch, a feature vector is extracted from a convolutional neural network model and concatenated at the viewing distance, for which the quality is predicted. The resulting vector is fed to fully connected layers to predict subjective scores for the considered viewing distance. The proposed method was evaluated using the Colourlab Image Database: Image Quality and Viewing Distance-changed Image Database. Both databases provide subjective scores at two different viewing distances. In the Colourlab Image Database: Image Quality we obtain a Pearson correlation of 0.87 at both 50 cm and 100 cm viewing distances, while in the Viewing Distance-changed Image Database we obtained a Pearson correlation of 0.93 and 0.94 at viewing distance of four and six times the image height. The results show the efficiency of our method and its generalization ability.
Collapse
|
41
|
Jayasankar U, Thirumal V, Ponnurangam D. A survey on data compression techniques: From the perspective of data quality, coding schemes, data type and applications. JOURNAL OF KING SAUD UNIVERSITY - COMPUTER AND INFORMATION SCIENCES 2021. [DOI: 10.1016/j.jksuci.2018.05.006] [Citation(s) in RCA: 55] [Impact Index Per Article: 13.8] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
42
|
Assessment of the Segmentation of RGB Remote Sensing Images: A Subjective Approach. REMOTE SENSING 2020. [DOI: 10.3390/rs12244152] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The evaluation of remote sensing imagery segmentation results plays an important role in the further image analysis and decision-making. The search for the optimal segmentation method for a particular data set and the suitability of segmentation results for the use in satellite image classification are examples where the proper image segmentation quality assessment can affect the quality of the final result. There is no extensive research related to the assessment of the segmentation effectiveness of the images. The designed objective quality assessment metrics that can be used to assess the quality of the obtained segmentation results usually take into account the subjective features of the human visual system (HVS). A novel approach is used in the article to estimate the effectiveness of satellite image segmentation by relating and determining the correlation between subjective and objective segmentation quality metrics. Pearson’s and Spearman’s correlation was used for satellite images after applying a k-means++ clustering algorithm based on colour information. Simultaneously, the dataset of the satellite images with ground truth (GT) based on the “DeepGlobe Land Cover Classification Challenge” dataset was constructed for testing three classes of quality metrics for satellite image segmentation.
Collapse
|
43
|
Full-Reference Quality Metric Based on Neural Network to Assess the Visual Quality of Remote Sensing Images. REMOTE SENSING 2020. [DOI: 10.3390/rs12152349] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Remote sensing images are subject to different types of degradations. The visual quality of such images is important because their visual inspection and analysis are still widely used in practice. To characterize the visual quality of remote sensing images, the use of specialized visual quality metrics is desired. Although the attempts to create such metrics are limited, there is a great number of visual quality metrics designed for other applications. Our idea is that some of these metrics can be employed in remote sensing under the condition that those metrics have been designed for the same distortion types. Thus, image databases that contain images with types of distortions that are of interest should be looked for. It has been checked what known visual quality metrics perform well for images with such degradations and an opportunity to design neural network-based combined metrics with improved performance has been studied. It is shown that for such combined metrics, their Spearman correlation coefficient with mean opinion score exceeds 0.97 for subsets of images in the Tampere Image Database (TID2013). Since different types of elementary metric pre-processing and neural network design have been considered, it has been demonstrated that it is enough to have two hidden layers and about twenty inputs. Examples of using known and designed visual quality metrics in remote sensing are presented.
Collapse
|
44
|
Li Y, Hu W, Zhang X, Xu Z, Ni J, Ligthart LP. Adaptive terahertz image super-resolution with adjustable convolutional neural network. OPTICS EXPRESS 2020; 28:22200-22217. [PMID: 32752486 DOI: 10.1364/oe.394943] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/14/2020] [Accepted: 07/07/2020] [Indexed: 06/11/2023]
Abstract
During the real-aperture-scanning imaging process, terahertz (THz) images are often plagued with the problem of low spatial resolution. Therefore, an accommodative super-resolution framework for THz images is proposed. Specifically, the 3D degradation model for the imaging system is firstly proposed by incorporating the focused THz beam distribution, which determines the relationship between the imaging range and the corresponding image restoration level. Secondly, an adjustable CNN is introduced to cope with this range dependent super-resolution problem. By simply tuning an interpolation parameter, the network can be adjusted to produce arbitrary restoration levels between the trained fixed levels without extra training. Finally, by selecting the appropriate interpolation coefficient according to the measured imaging range, each THz image can be coped with its matched network and reach the outstanding super-resolution effect. Both the simulated and real tested data, acquired by a 160 ∼ 220 GHz imager, have been used to demonstrate the superiority of our method.
Collapse
|
45
|
Ko H, Lee DY, Cho S, Bovik AC. Quality Prediction on Deep Generative Images. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; 29:5964-5979. [PMID: 32310772 DOI: 10.1109/tip.2020.2987180] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
In recent years, deep neural networks have been utilized in a wide variety of applications including image generation. In particular, generative adversarial networks (GANs) are able to produce highly realistic pictures as part of tasks such as image compression. As with standard compression, it is desirable to be able to automatically assess the perceptual quality of generative images to monitor and control the encode process. However, existing image quality algorithms are ineffective on GAN generated content, especially on textured regions and at high compressions. Here we propose a new "naturalness"-based image quality predictor for generative images. Our new GAN picture quality predictor is built using a multi-stage parallel boosting system based on structural similarity features and measurements of statistical similarity. To enable model development and testing, we also constructed a subjective GAN image quality database containing (distorted) GAN images and collected human opinions of them. Our experimental results indicate that our proposed GAN IQA model delivers superior quality predictions on the generative image datasets, as well as on traditional image quality datasets.
Collapse
|
46
|
Mason A, Rioux J, Clarke SE, Costa A, Schmidt M, Keough V, Huynh T, Beyea S. Comparison of Objective Image Quality Metrics to Expert Radiologists' Scoring of Diagnostic Quality of MR Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:1064-1072. [PMID: 31535985 DOI: 10.1109/tmi.2019.2930338] [Citation(s) in RCA: 68] [Impact Index Per Article: 13.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Image quality metrics (IQMs) such as root mean square error (RMSE) and structural similarity index (SSIM) are commonly used in the evaluation and optimization of accelerated magnetic resonance imaging (MRI) acquisition and reconstruction strategies. However, it is unknown how well these indices relate to a radiologist's perception of diagnostic image quality. In this study, we compare the image quality scores of five radiologists with the RMSE, SSIM, and other potentially useful IQMs: peak signal to noise ratio (PSNR) multi-scale SSIM (MSSSIM), information-weighted SSIM (IWSSIM), gradient magnitude similarity deviation (GMSD), feature similarity index (FSIM), high dynamic range visible difference predictor (HDRVDP), noise quality metric (NQM), and visual information fidelity (VIF). The comparison uses a database of MR images of the brain and abdomen that have been retrospectively degraded by noise, blurring, undersampling, motion, and wavelet compression for a total of 414 degraded images. A total of 1017 subjective scores were assigned by five radiologists. IQM performance was measured via the Spearman rank order correlation coefficient (SROCC) and statistically significant differences in the residuals of the IQM scores and radiologists' scores were tested. When considering SROCC calculated from combining scores from all radiologists across all image types, RMSE and SSIM had lower SROCC than six of the other IQMs included in the study (VIF, FSIM, NQM, GMSD, IWSSIM, and HDRVDP). In no case did SSIM have a higher SROCC or significantly smaller residuals than RMSE. These results should be considered when choosing an IQM in future imaging studies.
Collapse
|
47
|
Inverse Halftoning Methods Based on Deep Learning and Their Evaluation Metrics: A Review. APPLIED SCIENCES-BASEL 2020. [DOI: 10.3390/app10041521] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Inverse halftoning is an ill-posed problem that refers to the problem of restoring continuous-tone images from their halftone versions. Although much progress has been achieved over the last decades, the restored images still suffer from detail loss and visual artifacts. Recent studies show that inverse halftoning methods based on deep learning are superior to other traditional methods, and thus this paper aimed to systematically review the inverse halftone methods based on deep learning, so as to provide a reference for the development of inverse halftoning. In this paper, we firstly proposed a classification method for inverse halftoning methods on the basis of the source of halftone images. Then, two types of inverse halftoning methods for digital halftone images and scanned halftone images were investigated in terms of network architecture, loss functions, and training strategies. Furthermore, we studied existing image quality evaluation including subjective and objective evaluation by experiments. The evaluation results demonstrated that methods based on multiple subnetworks and methods based on multi-stage strategies are superior to other methods. In addition, the perceptual loss and the gradient loss are helpful for improving the quality of restored images. Finally, we gave the future research directions by analyzing the shortcomings of existing inverse halftoning methods.
Collapse
|
48
|
A Shift-Dependent Measure of Extended Cumulative Entropy and Its Applications in Blind Image Quality Assessment. Symmetry (Basel) 2020. [DOI: 10.3390/sym12020316] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Recently, Tahmasebi and Eskandarzadeh introduced a new extended cumulative entropy (ECE). In this paper, we present results on shift-dependent measure of ECE and its dynamic past version. These results contain stochastic order, upper and lower bounds, the symmetry property and some relationships with other reliability functions. We also discuss some properties of conditional weighted ECE under some assumptions. Finally, we propose a nonparametric estimator of this new measure and study its practical results in blind image quality assessment.
Collapse
|
49
|
Zhou W, Shi L, Chen Z, Zhang J. Tensor Oriented No-Reference Light Field Image Quality Assessment. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; 29:4070-4084. [PMID: 32012015 DOI: 10.1109/tip.2020.2969777] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Light field image (LFI) quality assessment is becoming more and more important, which helps to better guide the acquisition, processing and application of immersive media. However, due to the inherent high dimensional characteristics of LFI, the LFI quality assessment turns into a multi-dimensional problem that requires consideration of the quality degradation in both spatial and angular dimensions. Therefore, we propose a novel Tensor oriented No-reference Light Field image Quality evaluator (Tensor-NLFQ) based on tensor theory. Specifically, since the LFI is regarded as a low-rank 4D tensor, the principal components of four oriented sub-aperture view stacks are obtained via Tucker decomposition. Then, the Principal Component Spatial Characteristic (PCSC) is designed to measure the spatial-dimensional quality of LFI considering its global naturalness and local frequency properties. Finally, the Tensor Angular Variation Index (TAVI) is proposed to measure angular consistency quality by analyzing the structural similarity distribution between the first principal component and each view in the view stack. Extensive experimental results on four publicly available LFI quality databases demonstrate that the proposed Tensor-NLFQ model outperforms state-of-the-art 2D, 3D, multi-view, and LFI quality assessment algorithms.
Collapse
|
50
|
Kim W, Nguyen AD, Lee S, Bovik AC. Dynamic Receptive Field Generation for Full-Reference Image Quality Assessment. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; 29:4219-4231. [PMID: 31995494 DOI: 10.1109/tip.2020.2968283] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Most full-reference image quality assessment (FR-IQA) methods advanced to date have been holistically designed without regard to the type of distortion impairing the image. However, the perception of distortion depends nonlinearly on the distortion type. Here we propose a novel FR-IQA framework that dynamically generates receptive fields responsive to distortion type. Our proposed method-dynamic receptive field generation based image quality assessor (DRF-IQA)-separates the process of FR-IQA into two streams: 1) dynamic error representation and 2) visual sensitivity-based quality pooling. The first stream generates dynamic receptive fields on the input distorted image, implemented by a trained convolutional neural network (CNN), then the generated receptive field profiles are convolved with the distorted and reference images, and differenced to produce spatial error maps. In the second stream, a visual sensitivity map is generated. The visual sensitivity map is used to weight the spatial error map. The experimental results show that the proposed model achieves state-of-the-art prediction accuracy on various open IQA databases.
Collapse
|