1
|
Xia Y, Shi Y. Diffusion MRI harmonization via personalized template mapping. Hum Brain Mapp 2024; 45:e26661. [PMID: 38520363 PMCID: PMC10960558 DOI: 10.1002/hbm.26661] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2023] [Revised: 11/17/2023] [Accepted: 03/07/2024] [Indexed: 03/25/2024] Open
Abstract
One fundamental challenge in diffusion magnetic resonance imaging (dMRI) harmonization is to disentangle the contributions of scanner-related effects from the variable brain anatomy for the observed imaging signals. Conventional harmonization methods rely on establishing an atlas space to resolve anatomical variability and generate a unified inter-site mapping function. However, this approach is limited in accounting for the misalignment of neuroanatomy that still widely persists even after registration, especially in regions close to cortical boundaries. To overcome this challenge, we propose a personalized framework in this paper to more effectively address the confounding from the misalignment of neuroanatomy in dMRI harmonization. Instead of using a common template representing site-effects for all subjects, the main novelty of our method is the adaptive computation of personalized templates for both source and target scanning sites to estimate the inter-site mapping function. We integrate our method with the rotation invariant spherical harmonics (RISH) features to achieve the harmonization of dMRI signals between sites. In our experiments, the proposed approach is applied to harmonize the dMRI data acquired from two scanning platforms: Siemens Prisma and GE MR750 from the Adolescent Brain Cognitive Development dataset and compared with a state-of-the-art method based on RISH features. Our results indicate that the proposed harmonization framework achieves superior performance not only in reducing inter-site variations due to scanner differences but also in preserving sex-related biological variability in original cohorts. Moreover, we assess the impact of harmonization on the estimation of fiber orientation distributions and show the robustness of the personalized harmonization procedure in preserving the fiber orientation of original dMRI signals.
Collapse
Affiliation(s)
- Yihao Xia
- USC Stevens Neuroimaging and Informatics Institute, Keck School of MedicineUniversity of Southern CaliforniaLos AngelesCaliforniaUSA
- Department of Electrical and Computer Engineering, Viterbi School of EngineeringUniversity of Southern CaliforniaLos AngelesCaliforniaUSA
| | - Yonggang Shi
- USC Stevens Neuroimaging and Informatics Institute, Keck School of MedicineUniversity of Southern CaliforniaLos AngelesCaliforniaUSA
- Department of Electrical and Computer Engineering, Viterbi School of EngineeringUniversity of Southern CaliforniaLos AngelesCaliforniaUSA
| |
Collapse
|
2
|
Yao X, Cao Q, Feng X, Cheng G, Han J. Learning to Assess Image Quality Like an Observer. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:8324-8336. [PMID: 35196244 DOI: 10.1109/tnnls.2022.3149534] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Human observers are the ultimate receivers and evaluators of the image visual information and have powerful perception ability of visual quality with short-term global perception and long-term regional observation. Thus, it is natural to design an image quality assessment (IQA) computational model to act like an observer for accurately predicting the human perception of image quality. Inspired by this, here, we propose a novel observer-like network (OLN) to perform IQA by jointly considering the global glimpsing information and local scanning information. Specifically, the OLN consists of a global distortion perception (GDP) module and a local distortion observation (LDO) module. The GDP module is designed to mimic the observer's global perception of image quality through performing classification of images' distortion categories and levels. Simultaneously, to simulate the human local observation behavior, the LDO module attempts to gather the long-term regional observation information of the distorted images by continuously tracing the human scanpath in the observer-like scanning manner. By leveraging the bilinear pooling layer to collaborate the short-term global perception with the long-term regional observation, our network precisely predicts the quality scores of distorted images, such as human observers. Comprehensive experiments on the public datasets powerfully demonstrate that the proposed OLN achieves state-of-the-art performance.
Collapse
|
3
|
He W, Luo Z. Blind Quality Assessment of Images Containing Objects of Interest. SENSORS (BASEL, SWITZERLAND) 2023; 23:8205. [PMID: 37837037 PMCID: PMC10575444 DOI: 10.3390/s23198205] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/24/2023] [Revised: 08/28/2023] [Accepted: 09/27/2023] [Indexed: 10/15/2023]
Abstract
To monitor objects of interest, such as wildlife and people, image-capturing devices are used to collect a large number of images with and without objects of interest. As we are recording valuable information about the behavior and activity of objects, the quality of images containing objects of interest should be better than that of images without objects of interest, even if the former exhibits more severe distortion than the latter. However, according to current methods, quality assessments produce the opposite results. In this study, we propose an end-to-end model, named DETR-IQA (detection transformer image quality assessment), which extends the capability to perform object detection and blind image quality assessment (IQA) simultaneously by adding IQA heads comprising simple multi-layer perceptrons at the top of the DETRs (detection transformers) decoder. Using IQA heads, DETR-IQA carried out blind IQAs based on the weighted fusion of the distortion degree of the region of objects of interest and the other regions of the image; the predicted quality score of images containing objects of interest was generally greater than that of images without objects of interest. Currently, the subjective quality score of all public datasets is in accordance with the distortion of images and does not consider objects of interest. We manually extracted the images in which the five predefined classes of objects were the main contents of the largest authentic distortion dataset, KonIQ-10k, which was used as the experimental dataset. The experimental results show that with slight degradation in object detection performance and simple IQA heads, the values of PLCC and SRCC were 0.785 and 0.727, respectively, and exceeded those of some deep learning-based IQA models that are specially designed for only performing IQA. With the negligible increase in the computation and complexity of object detection and without a decrease in inference speeds, DETR-IQA can perform object detection and IQA via multi-tasking and substantially reduce the workload.
Collapse
Affiliation(s)
- Wentong He
- Computer Network Information Center, Chinese Academy of Sciences, Beijing 100190, China;
- University of Chinese Academy of Sciences, Beijing 100049, China
| | - Ze Luo
- Computer Network Information Center, Chinese Academy of Sciences, Beijing 100190, China;
| |
Collapse
|
4
|
Han Z, Liu Y, Xie R, Zhai G. Image Quality Assessment for Realistic Zoom Photos. SENSORS (BASEL, SWITZERLAND) 2023; 23:4724. [PMID: 37430638 DOI: 10.3390/s23104724] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/25/2023] [Revised: 04/25/2023] [Accepted: 05/08/2023] [Indexed: 07/12/2023]
Abstract
New CMOS imaging sensor (CIS) techniques in smartphones have helped user-generated content dominate our lives over traditional DSLRs. However, tiny sensor sizes and fixed focal lengths also lead to more grainy details, especially for zoom photos. Moreover, multi-frame stacking and post-sharpening algorithms would produce zigzag textures and over-sharpened appearances, for which traditional image-quality metrics may over-estimate. To solve this problem, a real-world zoom photo database is first constructed in this paper, which includes 900 tele-photos from 20 different mobile sensors and ISPs. Then we propose a novel no-reference zoom quality metric which incorporates the traditional estimation of sharpness and the concept of image naturalness. More specifically, for the measurement of image sharpness, we are the first to combine the total energy of the predicted gradient image with the entropy of the residual term under the framework of free-energy theory. To further compensate for the influence of over-sharpening effect and other artifacts, a set of model parameters of mean subtracted contrast normalized (MSCN) coefficients are utilized as the natural statistics representatives. Finally, these two measures are combined linearly. Experimental results on the zoom photo database demonstrate that our quality metric can achieve SROCC and PLCC over 0.91, while the performance of single sharpness or naturalness index is around 0.85. Moreover, compared with the best tested general-purpose and sharpness models, our zoom metric outperforms them by 0.072 and 0.064 in SROCC, respectively.
Collapse
Affiliation(s)
- Zongxi Han
- Institute of Image Communication and Information Processing, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Yutao Liu
- School of Computer Science and Technology, Ocean University of China, Qingdao 266100, China
| | - Rong Xie
- Institute of Image Communication and Information Processing, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Guangtao Zhai
- Institute of Image Communication and Information Processing, Shanghai Jiao Tong University, Shanghai 200240, China
| |
Collapse
|
5
|
Li L, Zhong C, He Z. Blind quality assessment of multi-exposure fused images considering the detail, structure and color characteristics. PLoS One 2023; 18:e0283096. [PMID: 37023106 PMCID: PMC10079045 DOI: 10.1371/journal.pone.0283096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2022] [Accepted: 03/01/2023] [Indexed: 04/07/2023] Open
Abstract
In the process of multi-exposure image fusion (MEF), the appearance of various distortions will inevitably cause the deterioration of visual quality. It is essential to predict the visual quality of MEF images. In this work, a novel blind image quality assessment (IQA) method is proposed for MEF images considering the detail, structure, and color characteristics. Specifically, to better perceive the detail and structure distortion, based on the joint bilateral filtering, the MEF image is decomposed into two layers (i.e., the energy layer and the structure layer). Obviously, this is a symmetric process that the two decomposition results can independently and almost completely describe the information of MEF images. As the former layer contains rich intensity information and the latter captures some image structures, some energy-related and structure-related features are extracted from these two layers to perceive the detail and structure distortion phenomena. Besides, some color-related features are also obtained to present the color degradation which are combined with the above energy-related and structure-related features for quality regression. Experimental results on the public MEF image database demonstrate that the proposed method achieves higher performance than the state-of-the-art quality assessment ones.
Collapse
Affiliation(s)
- Lijun Li
- School of Automation, Southeast University, Nanjing, China
| | - Caiming Zhong
- College of Science and Technology, Ningbo University, Ningbo, China
| | - Zhouyan He
- College of Science and Technology, Ningbo University, Ningbo, China
- School of Automation, Qingdao University, Qingdao, China
| |
Collapse
|
6
|
Two-channel deep recursive multi-scale network based on multi-attention for no-reference image quality assessment. INT J MACH LEARN CYB 2023. [DOI: 10.1007/s13042-023-01773-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
7
|
Hybrid No-Reference Quality Assessment for Surveillance Images. INFORMATION 2022. [DOI: 10.3390/info13120588] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/23/2022] Open
Abstract
Intelligent video surveillance (IVS) technology is widely used in various security systems. However, quality degradation in surveillance images (SIs) may affect its performance on vision-based tasks, leading to the difficulties in the IVS system extracting valid information from SIs. In this paper, we propose a hybrid no-reference image quality assessment (NR IQA) model for SIs that can help to identify undesired distortions and provide useful guidelines for IVS technology. Specifically, we first extract two main types of quality-aware features: the low-level visual features related to various distortions, and the high-level semantic information, which is extracted by a state-of-the-art (SOTA) vision transformer backbone. Then, we fuse these two kinds of features into the final quality-aware feature vector, which is mapped into the quality index through the feature regression module. Our experimental results on two surveillance content quality databases demonstrate that the proposed model achieves the best performance compared to the SOTA on NR IQA metrics.
Collapse
|
8
|
Varga D. No-Reference Video Quality Assessment Using the Temporal Statistics of Global and Local Image Features. SENSORS (BASEL, SWITZERLAND) 2022; 22:9696. [PMID: 36560065 PMCID: PMC9780801 DOI: 10.3390/s22249696] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/30/2022] [Revised: 11/27/2022] [Accepted: 12/09/2022] [Indexed: 06/17/2023]
Abstract
During acquisition, storage, and transmission, the quality of digital videos degrades significantly. Low-quality videos lead to the failure of many computer vision applications, such as object tracking or detection, intelligent surveillance, etc. Over the years, many different features have been developed to resolve the problem of no-reference video quality assessment (NR-VQA). In this paper, we propose a novel NR-VQA algorithm that integrates the fusion of temporal statistics of local and global image features with an ensemble learning framework in a single architecture. Namely, the temporal statistics of global features reflect all parts of the video frames, while the temporal statistics of local features reflect the details. Specifically, we apply a broad spectrum of statistics of local and global features to characterize the variety of possible video distortions. In order to study the effectiveness of the method introduced in this paper, we conducted experiments on two large benchmark databases, i.e., KoNViD-1k and LIVE VQC, which contain authentic distortions, and we compared it to 14 other well-known NR-VQA algorithms. The experimental results show that the proposed method is able to achieve greatly improved results on the considered benchmark datasets. Namely, the proposed method exhibits significant progress in performance over other recent NR-VQA approaches.
Collapse
|
9
|
No-reference image quality assessment with multi-scale weighted residuals and channel attention mechanism. Soft comput 2022. [DOI: 10.1007/s00500-022-07535-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/10/2022]
|
10
|
Chen B, Zhu L, Kong C, Zhu H, Wang S, Li Z. No-Reference Image Quality Assessment by Hallucinating Pristine Features. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:6139-6151. [PMID: 36112560 DOI: 10.1109/tip.2022.3205770] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
In this paper, we propose a no-reference (NR) image quality assessment (IQA) method via feature level pseudo-reference (PR) hallucination. The proposed quality assessment framework is rooted in the view that the perceptually meaningful features could be well exploited to characterize the visual quality, and the natural image statistical behaviors are exploited in an effort to deliver the accurate predictions. Herein, the PR features from the distorted images are learned by a mutual learning scheme with the pristine reference as the supervision, and the discriminative characteristics of PR features are further ensured with the triplet constraints. Given a distorted image for quality inference, the feature level disentanglement is performed with an invertible neural layer for final quality prediction, leading to the PR and the corresponding distortion features for comparison. The effectiveness of our proposed method is demonstrated on four popular IQA databases, and superior performance on cross-database evaluation also reveals the high generalization capability of our method. The implementation of our method is publicly available on https://github.com/Baoliang93/FPR.
Collapse
|
11
|
Shi H, Wang L, Wang G. Blind Quality Prediction for View Synthesis Based on Heterogeneous Distortion Perception. SENSORS (BASEL, SWITZERLAND) 2022; 22:7081. [PMID: 36146438 PMCID: PMC9504726 DOI: 10.3390/s22187081] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/14/2022] [Revised: 09/07/2022] [Accepted: 09/08/2022] [Indexed: 06/16/2023]
Abstract
The quality of synthesized images directly affects the practical application of virtual view synthesis technology, which typically uses a depth-image-based rendering (DIBR) algorithm to generate a new viewpoint based on texture and depth images. Current view synthesis quality metrics commonly evaluate the quality of DIBR-synthesized images, where the DIBR process is computationally expensive and time-consuming. In addition, the existing view synthesis quality metrics cannot achieve robustness due to the shallow hand-crafted features. To avoid the complicated DIBR process and learn more efficient features, this paper presents a blind quality prediction model for view synthesis based on HEterogeneous DIstortion Perception, dubbed HEDIP, which predicts the image quality of view synthesis from texture and depth images. Specifically, the texture and depth images are first fused based on discrete cosine transform to simulate the distortion of view synthesis images, and then the spatial and gradient domain features are extracted in a Two-Channel Convolutional Neural Network (TCCNN). Finally, a fully connected layer maps the extracted features to a quality score. Notably, the ground-truth score of the source image cannot effectively represent the labels of each image patch during training due to the presence of local distortions in view synthesis image. So, we design a Heterogeneous Distortion Perception (HDP) module to provide effective training labels for each image patch. Experiments show that with the help of the HDP module, the proposed model can effectively predict the quality of view synthesis. Experimental results demonstrate the effectiveness of the proposed model.
Collapse
Affiliation(s)
- Haozhi Shi
- School of Physics, Xidian University, Xi’an 710071, China
| | - Lanmei Wang
- School of Physics, Xidian University, Xi’an 710071, China
| | - Guibao Wang
- School of Physics and Telecommunication Engineering, Shaanxi University of Technology, Hanzhong 723001, China
| |
Collapse
|
12
|
Varga D. A Human Visual System Inspired No-Reference Image Quality Assessment Method Based on Local Feature Descriptors. SENSORS (BASEL, SWITZERLAND) 2022; 22:6775. [PMID: 36146123 PMCID: PMC9502000 DOI: 10.3390/s22186775] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/27/2022] [Revised: 09/01/2022] [Accepted: 09/05/2022] [Indexed: 06/16/2023]
Abstract
Objective quality assessment of natural images plays a key role in many fields related to imaging and sensor technology. Thus, this paper intends to introduce an innovative quality-aware feature extraction method for no-reference image quality assessment (NR-IQA). To be more specific, a various sequence of HVS inspired filters were applied to the color channels of an input image to enhance those statistical regularities in the image to which the human visual system is sensitive. From the obtained feature maps, the statistics of a wide range of local feature descriptors were extracted to compile quality-aware features since they treat images from the human visual system's point of view. To prove the efficiency of the proposed method, it was compared to 16 state-of-the-art NR-IQA techniques on five large benchmark databases, i.e., CLIVE, KonIQ-10k, SPAQ, TID2013, and KADID-10k. It was demonstrated that the proposed method is superior to the state-of-the-art in terms of three different performance indices.
Collapse
|
13
|
Wang F, Chen J, Xie Z, Ai Y, Zhang W. Local sharpness failure detection of camera module lens based on image blur assessment. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03948-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
14
|
Zheng Y, Chen W, Lin R, Zhao T, Le Callet P. UIF: An Objective Quality Assessment for Underwater Image Enhancement. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:5456-5468. [PMID: 35951566 DOI: 10.1109/tip.2022.3196815] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Due to complex and volatile lighting environment, underwater imaging can be readily impaired by light scattering, warping, and noises. To improve the visual quality, Underwater Image Enhancement (UIE) techniques have been widely studied. Recent efforts have also been contributed to evaluate and compare the UIE performances with subjective and objective methods. However, the subjective evaluation is time-consuming and uneconomic for all images, while existing objective methods have limited capabilities for the newly-developed UIE approaches based on deep learning. To fill this gap, we propose an Underwater Image Fidelity (UIF) metric for objective evaluation of enhanced underwater images. By exploiting the statistical features of these images in CIELab space, we present the naturalness, sharpness, and structure indexes. Among them, the naturalness and sharpness indexes represent the visual improvements of enhanced images; the structure index indicates the structural similarity between the underwater images before and after UIE. We combine all indexes with a saliency-based spatial pooling and thus obtain the final UIF metric. To evaluate the proposed metric, we also establish a first-of-its-kind large-scale UIE database with subjective scores, namely Underwater Image Enhancement Database (UIED). Experimental results confirm that the proposed UIF metric outperforms a variety of underwater and general-purpose image quality metrics. The database and source code are available at https://github.com/z21110008/UIF.
Collapse
|
15
|
Li A, Wu J, Tian S, Li L, Dong W, Shi G. Blind image quality assessment based on progressive multi-task learning. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.05.043] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
16
|
Blind image quality assessment of magnetic resonance images with statistics of local intensity extrema. Inf Sci (N Y) 2022. [DOI: 10.1016/j.ins.2022.05.061] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/18/2022]
|
17
|
Cao J, Wu W, Wang R, Kwong S. No-reference image quality assessment by using convolutional neural networks via object detection. INT J MACH LEARN CYB 2022. [DOI: 10.1007/s13042-022-01611-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/16/2022]
|
18
|
Super Resolution Image Visual Quality Assessment Based on Feature Optimization. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:1263348. [PMID: 35769272 PMCID: PMC9236850 DOI: 10.1155/2022/1263348] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/12/2022] [Revised: 05/16/2022] [Accepted: 05/23/2022] [Indexed: 11/23/2022]
Abstract
Most existing no-referenced image quality assessment (NR-IQA) algorithms need to extract features first and then predict image quality. However, only a small number of features work in the model, and the rest will degrade the model performance. Consequently, an NR-IQA framework based on feature optimization is proposed to solve this problem and apply to the SR-IQA field. In this study, we designed a feature engineering method to solve this problem. Specifically, the features associate with the SR images were first collected and aggregated. Furthermore, several advanced feature selection algorithms were used to sort the feature sets according to their importance, and the importance matrix of features is obtained. Then, we examined the linear relationship between the number of features and Pearson linear correlation coefficient (PLCC) to determine the optimal number of features and the optimal feature selection algorithm, so as to obtain the optimal model. The results showed that the image quality scores predicted by the optimal model are in good agreement with the human subjective scores. Adopting the proposed feature optimization framework, we can effectively reduce the number of features in the model and obtain better performance. The experimental results indicated that SR image quality can be accurately predicted using only a small part of image features. In summary, we proposed a feature optimization framework to solve the current problem of irrelevant features in SR-IQA, and an SR image quality assessment model was proposed consequently.
Collapse
|
19
|
No-Reference Quality Assessment of Authentically Distorted Images Based on Local and Global Features. J Imaging 2022; 8:jimaging8060173. [PMID: 35735972 PMCID: PMC9224559 DOI: 10.3390/jimaging8060173] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2022] [Revised: 06/15/2022] [Accepted: 06/17/2022] [Indexed: 01/25/2023] Open
Abstract
With the development of digital imaging techniques, image quality assessment methods are receiving more attention in the literature. Since distortion-free versions of camera images in many practical, everyday applications are not available, the need for effective no-reference image quality assessment algorithms is growing. Therefore, this paper introduces a novel no-reference image quality assessment algorithm for the objective evaluation of authentically distorted images. Specifically, we apply a broad spectrum of local and global feature vectors to characterize the variety of authentic distortions. Among the employed local features, the statistics of popular local feature descriptors, such as SURF, FAST, BRISK, or KAZE, are proposed for NR-IQA; other features are also introduced to boost the performances of local features. The proposed method was compared to 12 other state-of-the-art algorithms on popular and accepted benchmark datasets containing RGB images with authentic distortions (CLIVE, KonIQ-10k, and SPAQ). The introduced algorithm significantly outperforms the state-of-the-art in terms of correlation with human perceptual quality ratings.
Collapse
|
20
|
Khan MU, Luo MR, Tian D. No-reference image quality metrics for color domain modified images. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2022; 39:B65-B77. [PMID: 36215544 DOI: 10.1364/josaa.450595] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Accepted: 05/16/2022] [Indexed: 06/16/2023]
Abstract
Predicting the quality of natural images without using a reference image has always been a challenging task. Numerous approaches have been proposed in the past, but they mainly focused on spatial and frequency domain degradations like blur, noise, and compression. Image quality metrics (IQMs) in literature perform with quite a high accuracy for such types of degraded images. However, their performances are not good on the images modified in the color domain. In this study, psychophysical experiments were conducted to assess the quality of the color domain images. A new dataset was developed for this purpose. Additionally, a second dataset consisting of color domain modified images from the three previously published datasets were used in the psychophysical experiments. The newly developed dataset was then used to develop three IQMs based on absolute values, relative values, and statistical analysis of image color appearance attributes. Their performances were then evaluated together with five spatial domain IQMs from the literature using cross-database evaluation methodology. The results showed that the color-domain IQMs outperformed the other models. The absolute and relative attributes-based models, when combined, achieved the best performance. The present results suggest that more effort is needed to improve the performance of color domain IQMs for image quality estimation.
Collapse
|
21
|
Yang W, Wu J, Tian S, Li L, Dong W, Shi G. Fine-Grained Image Quality Caption With Hierarchical Semantics Degradation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:3578-3590. [PMID: 35511851 DOI: 10.1109/tip.2022.3171445] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Blind image quality assessment (BIQA), which is capable of precisely and automatically estimating human perceived image quality with no pristine image for comparison, attracts extensive attention and is of wide applications. Recently, many existing BIQA methods commonly represent image quality with a quantitative value, which is inconsistent with human cognition. Generally, human beings are good at perceiving image quality in terms of semantic description rather than quantitative value. Moreover, cognition is a needs-oriented task where humans are able to extract image contents with local to global semantics as they need. The mediocre quality value represents coarse or holistic image quality and fails to reflect degradation on hierarchical semantics. In this paper, to comply with human cognition, a novel quality caption model is inventively proposed to measure fine-grained image quality with hierarchical semantics degradation. Research on human visual system indicates there are hierarchy and reverse hierarchy correlations between hierarchical semantics. Meanwhile, empirical evidence shows that there are also bi-directional degradation dependencies between them. Thus, a novel bi-directional relationship-based network (BDRNet) is proposed for semantics degradation description, through adaptively exploring those correlations and degradation dependencies in a bi-directional manner. Extensive experiments demonstrate that our method outperforms the state-of-the-arts in terms of both evaluation performance and generalization ability.
Collapse
|
22
|
Two Low-Level Feature Distributions Based No Reference Image Quality Assessment. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12104975] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
No reference image quality assessment (NR IQA) aims to develop quantitative measures to automatically and accurately estimate perceptual image quality without any prior information about the reference image. In this paper, we introduce two low-level feature distributions (TLLFD) based method for NR IQA. Different from the deep learning method, the proposed method characterizes image quality with the distributions of low-level features, thus it has few parameters, simple model, high efficiency, and strong robustness. First, the texture change of distorted image is extracted by the weighted histogram of generalized local binary pattern. Second, the Weibull distribution of gradient is extracted to represent the structural change of the distorted image. Furthermore, support vector regression is adopted to model the complex nonlinear relationship between feature space and quality measure. Finally, numerical tests are performed on LIVE, CISQ, MICT, and TID2008 standard databases for five different distortion categories JPEG2000 (JP2K), JPEG, White Noise (WN), Gaussian Blur (GB), and Fast Fading (FF). The experimental results indicate that TLLFD method achieves superior performance and strong generalization for image quality prediction as compared to state-of-the-art full-reference, no reference, and even deep learning IQA methods.
Collapse
|
23
|
Yang M, Yin G, Wang H, Dong J, Xie Z, Zheng B. A Underwater Sequence Image Dataset for Sharpness and Color Analysis. SENSORS 2022; 22:s22093550. [PMID: 35591240 PMCID: PMC9100472 DOI: 10.3390/s22093550] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/15/2022] [Revised: 04/27/2022] [Accepted: 04/29/2022] [Indexed: 02/01/2023]
Abstract
The complex underwater environment usually leads to the problem of quality degradation in underwater images, and the distortion of sharpness and color are the main factors to the quality of underwater images. The paper discloses an underwater sequence image dataset called TankImage-I with gradually changing sharpness and color distortion collected in a pool. TankImage-I contains two plane targets, a total of 78 images. It includes two lighting conditions and three different water transparency. The imaging distance is also changed during the photographing process. The paper introduces the relevant details of the photographing process, and provides the measurement results of the sharpness and color distortion of the sequence images. In addition, we verify the performance of 13 image quality assessment methods on TankImage-I, and analyze the results of 13 image quality assessment methods from the aspects of sharpness and color, which provides a reference for the design and improvement of underwater image quality assessment algorithm and underwater imaging system design.
Collapse
Affiliation(s)
- Miao Yang
- School of Electronic Engineering, Jiangsu Ocean University, Lianyungang 222005, China; (M.Y.); (H.W.); (J.D.); (Z.X.)
| | - Ge Yin
- School of Electronic Engineering, Jiangsu Ocean University, Lianyungang 222005, China; (M.Y.); (H.W.); (J.D.); (Z.X.)
- Correspondence:
| | - Haiwen Wang
- School of Electronic Engineering, Jiangsu Ocean University, Lianyungang 222005, China; (M.Y.); (H.W.); (J.D.); (Z.X.)
| | - Jinnai Dong
- School of Electronic Engineering, Jiangsu Ocean University, Lianyungang 222005, China; (M.Y.); (H.W.); (J.D.); (Z.X.)
| | - Zhuoran Xie
- School of Electronic Engineering, Jiangsu Ocean University, Lianyungang 222005, China; (M.Y.); (H.W.); (J.D.); (Z.X.)
| | - Bing Zheng
- Department of Information Science and Engineering, Ocean University of China, Qingdao 266001, China;
| |
Collapse
|
24
|
Dong X, Fu L, Liu Q. No-reference image quality assessment for confocal endoscopy images with perceptual local descriptor. JOURNAL OF BIOMEDICAL OPTICS 2022; 27:056503. [PMID: 35585672 PMCID: PMC9116465 DOI: 10.1117/1.jbo.27.5.056503] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/11/2021] [Accepted: 04/29/2022] [Indexed: 06/15/2023]
Abstract
SIGNIFICANCE Confocal endoscopy images often suffer distortions, resulting in image quality degradation and information loss, increasing the difficulty of diagnosis and even leading to misdiagnosis. It is important to assess image quality and filter images with low diagnostic value before diagnosis. AIM We propose a no-reference image quality assessment (IQA) method for confocal endoscopy images based on Weber's law and local descriptors. The proposed method can detect the severity of image degradation by capturing the perceptual structure of an image. APPROACH We created a new dataset of 642 confocal endoscopy images to validate the performance of the proposed method. We then conducted extensive experiments to compare the accuracy and speed of the proposed method with other state-of-the-art IQA methods. RESULTS Experimental results demonstrate that the proposed method achieved an SROCC of 0.85 and outperformed other IQA methods. CONCLUSIONS Given its high consistency in subjective quality assessment, the proposed method can screen high-quality images in practical applications and contribute to diagnosis.
Collapse
Affiliation(s)
- Xiangjiang Dong
- Huazhong University of Science and Technology, Wuhan National Laboratory for Optoelectronics, Wuhan, China
| | - Ling Fu
- Huazhong University of Science and Technology, Wuhan National Laboratory for Optoelectronics, Wuhan, China
- Hainan University, School of Biomedical Engineering, Key Laboratory of Biomedical Engineering of Hainan Province, Hainan, China
| | - Qian Liu
- Hainan University, School of Biomedical Engineering, Key Laboratory of Biomedical Engineering of Hainan Province, Hainan, China
| |
Collapse
|
25
|
Haze Level Evaluation Using Dark and Bright Channel Prior Information. ATMOSPHERE 2022. [DOI: 10.3390/atmos13050683] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Haze level evaluation is highly desired in outdoor scene monitoring applications. However, there are relatively few approaches available in this area. In this paper, a novel haze level evaluation strategy for real-world outdoor scenes is presented. The idea is inspired by the utilization of dark and bright channel prior (DBCP) for haze removal. The change between hazy and haze-free scenes in bright channels could serve as a haze level indicator, and we have named it DBCP-I. The variation of contrast between dark and bright channels in a single hazy image also contains useful information to reflect haze level. By searching for a segmentation threshold, a metric called DBCP-II is proposed. Combining the strengths of the above two indicators, a hybrid metric named DBCP-III is constructed to achieve better performance. The experiment results on public, real-world benchmark datasets show the advantages of the proposed methods in terms of assessment accuracy with subjective human ratings. The study is first-of-its-kind with preliminary exploration in the field of haze level evaluation for real outdoor scenes, and it has a great potential to promote research in autonomous driving and automatic air quality monitoring. The open-source codes of the proposed algorithms are free to download.
Collapse
|
26
|
Tang T, Li L, Wu X, Chen R, Li H, Lu G, Cheng L. TSA-SCC: Text Semantic-Aware Screen Content Coding With Ultra Low Bitrate. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:2463-2477. [PMID: 35196232 DOI: 10.1109/tip.2022.3152003] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Due to the rapid growth of web conferences, remote screen sharing, and online games, screen content has become an important type of internet media information and over 90% of online media interactions are screen based. Meanwhile, as the main component in the screen content, textual information averagely takes up over 40% of the whole image on various commonly used screen content datasets. However, it is difficult to compress the textual information by using the traditional coding schemes as HEVC, which assumes strong spatial and temporal correlations within the image/video. State-of-the-art screen content coding (SCC) standard as HEVC-SCC still adopts a block-based coding framework and does not consider the text semantics for compression, thus inevitably blurring texts at a lower bitrate. In this paper, we propose a general text semantic-aware screen content coding scheme (TSA-SCC) for ultra low bitrate setting. This method detects the abrupt picture in a screen content video (or image), recognizes textual information (including word, position, font type, font size and font color) in the abrupt picture based on neural networks, and encodes texts with text coding tools. The other pictures as well as the background image after removing texts from the abrupt picture via inpainting, are encoded with HEVC-SCC. Compared with HEVC-SCC, the proposed method TSA-SCC reduces bitrate by up to 3× at a similar compression quality. Moreover, TSA-SCC achieves much better visual quality with less bitrate consumption when encoding the screen content video/image at ultra low bitrates.
Collapse
|
27
|
Jiang Q, Liu Z, Gu K, Shao F, Zhang X, Liu H, Lin W. Single Image Super-Resolution Quality Assessment: A Real-World Dataset, Subjective Studies, and an Objective Metric. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:2279-2294. [PMID: 35239481 DOI: 10.1109/tip.2022.3154588] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Numerous single image super-resolution (SISR) algorithms have been proposed during the past years to reconstruct a high-resolution (HR) image from its low-resolution (LR) observation. However, how to fairly compare the performance of different SISR algorithms/results remains a challenging problem. So far, the lack of comprehensive human subjective study on large-scale real-world SISR datasets and accurate objective SISR quality assessment metrics makes it unreliable to truly understand the performance of different SISR algorithms. We in this paper make efforts to tackle these two issues. Firstly, we construct a real-world SISR quality dataset (i.e., RealSRQ) and conduct human subjective studies to compare the performance of the representative SISR algorithms. Secondly, we propose a new objective metric, i.e., KLTSRQA, based on the Karhunen-Loéve Transform (KLT) to evaluate the quality of SISR images in a no-reference (NR) manner. Experiments on our constructed RealSRQ and the latest synthetic SISR quality dataset (i.e., QADS) have demonstrated the superiority of our proposed KLTSRQA metric, achieving higher consistency with human subjective scores than relevant existing NR image quality assessment (NR-IQA) metrics. The dataset and the code will be made available at https://github.com/Zhentao-Liu/RealSRQ-KLTSRQA.
Collapse
|
28
|
Subjective and Objective Quality Evaluation for Underwater Image Enhancement and Restoration. Symmetry (Basel) 2022. [DOI: 10.3390/sym14030558] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/04/2023] Open
Abstract
Since underwater imaging is affected by the complex water environment, it often leads to severe distortion of the underwater image. To improve the quality of underwater images, underwater image enhancement and restoration methods have been proposed. However, many underwater image enhancement and restoration methods produce over-enhancement or under-enhancement, which affects their application. To better design underwater image enhancement and restoration methods, it is necessary to research the underwater image quality evaluation (UIQE) for underwater image enhancement and restoration methods. Therefore, a subjective evaluation dataset for an underwater image enhancement and restoration method is constructed, and on this basis, an objective quality evaluation method of underwater images, based on the relative symmetry of underwater dark channel prior (UDCP) and the underwater bright channel prior (UBCP) is proposed. Specifically, considering underwater image enhancement in different scenarios, a UIQE dataset is constructed, which contains 405 underwater images, generated from 45 different underwater real images, using 9 representative underwater image enhancement methods. Then, a subjective quality evaluation of the UIQE database is studied. To quantitatively measure the quality of the enhanced and restored underwater images with different characteristics, an objective UIQE index (UIQEI) is used, by extracting and fusing four groups of features, including: (1) the joint statistics of normalized gradient magnitude (GM) and Laplacian of Gaussian (LOG) features, based on the underwater dark channel map; (2) the joint statistics of normalized gradient magnitude (GM) and Laplacian of Gaussian (LOG) features, based on the underwater bright channel map; (3) the saturation and colorfulness features; (4) the fog density feature; (5) the global contrast feature; these features capture key aspects of underwater images. Finally, the experimental results are analyzed, qualitatively and quantitatively, to illustrate the effectiveness of the proposed UIQEI method.
Collapse
|
29
|
No-Reference Image Quality Assessment Based on Image Multi-Scale Contour Prediction. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12062833] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Accurately assessing image quality is a challenging task, especially without a reference image. Currently, most of the no-reference image quality assessment methods still require reference images in the training stage, but reference images are usually not available in real scenes. In this paper, we proposed a model named MSIQA inspired by biological vision and a convolution neural network (CNN), which does not require reference images in the training and testing phases. The model contains two modules, a multi-scale contour prediction network that simulates the contour response of the human optic nerve to images at different distances, and a central attention peripheral inhibition module inspired by the receptive field mechanism of retinal ganglion cells. There are two steps in the training stage. In the first step, the multi-scale contour prediction network learns to predict the contour features of images in different scales, and in the second step, the model combines the central attention peripheral inhibition module to learn to predict the quality score of the image. In the experiments, our method has achieved excellent performance. The Pearson linear correlation coefficient of the MSIQA model test on the LIVE database reached 0.988.
Collapse
|
30
|
BMEFIQA: Blind Quality Assessment of Multi-Exposure Fused Images Based on Several Characteristics. ENTROPY 2022; 24:e24020285. [PMID: 35205579 PMCID: PMC8871194 DOI: 10.3390/e24020285] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/20/2021] [Revised: 02/05/2022] [Accepted: 02/08/2022] [Indexed: 12/05/2022]
Abstract
A multi-exposure fused (MEF) image is generated by multiple images with different exposure levels, but the transformation process will inevitably introduce various distortions. Therefore, it is worth discussing how to evaluate the visual quality of MEF images. This paper proposes a new blind quality assessment method for MEF images by considering their characteristics, and it is dubbed as BMEFIQA. More specifically, multiple features that represent different image attributes are extracted to perceive the various distortions of MEF images. Among them, structural, naturalness, and colorfulness features are utilized to describe the phenomena of structure destruction, unnatural presentation, and color distortion, respectively. All the captured features constitute a final feature vector for quality regression via random forest. Experimental results on a publicly available database show the superiority of the proposed BMEFIQA method to several blind quality assessment methods.
Collapse
|
31
|
Combined No-Reference Image Quality Metrics for Visual Quality Assessment Optimized for Remote Sensing Images. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12041986] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/10/2022]
Abstract
No-reference image quality assessment is one of the most demanding areas of image analysis for many applications where the results of the analysis should be strongly correlated with the quality of an input image and the corresponding reference image is unavailable. One of the examples might be remote sensing since the transmission of such obtained images often requires the use of lossy compression and they are often distorted, e.g., by the presence of noise and blur. Since the practical usefulness of acquired and/or preprocessed images is directly related to their quality, there is a need for the development of reliable and adequate no-reference metrics that do not need any reference images. As the performance and universality of many existing metrics are quite limited, one of the possible solutions is the design and application of combined metrics. Several possible approaches to their composition have been previously proposed and successfully used for full-reference metrics. In the paper, three possible approaches to the development and optimization of no-reference combined metrics are investigated and verified for the dataset of images containing distortions typical for remote sensing. The proposed approach leads to good results, significantly improving the correlation of the obtained results with subjective quality scores.
Collapse
|
32
|
|
33
|
A novel hybrid approach of ABC with SCA for the parameter optimization of SVR in blind image quality assessment. Neural Comput Appl 2022. [DOI: 10.1007/s00521-021-06435-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
34
|
Lu W, Sun W, Min X, Zhang Z, Wang T, Zhu W, Yang X, Zhai G. Blind Surveillance Image Quality Assessment via Deep Neural Network Combined with the Visual Saliency. ARTIF INTELL 2022. [DOI: 10.1007/978-3-031-20500-2_11] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/02/2023]
|
35
|
Cui Y, Jiang G, Yu M, Song Y. Local Visual and Global Deep Features Based Blind Stitched Panoramic Image Quality Evaluation Using Ensemble Learning. IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE 2022. [DOI: 10.1109/tetci.2022.3165935] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
36
|
Takam Tchendjou G, Simeu E. Visual Perceptual Quality Assessment Based on Blind Machine Learning Techniques. SENSORS (BASEL, SWITZERLAND) 2021; 22:s22010175. [PMID: 35009718 PMCID: PMC8749612 DOI: 10.3390/s22010175] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/25/2021] [Revised: 12/14/2021] [Accepted: 12/18/2021] [Indexed: 05/03/2023]
Abstract
This paper presents the construction of a new objective method for estimation of visual perceiving quality. The proposal provides an assessment of image quality without the need for a reference image or a specific distortion assumption. Two main processes have been used to build our models: The first one uses deep learning with a convolutional neural network process, without any preprocessing. The second objective visual quality is computed by pooling several image features extracted from different concepts: the natural scene statistic in the spatial domain, the gradient magnitude, the Laplacian of Gaussian, as well as the spectral and spatial entropies. The features extracted from the image file are used as the input of machine learning techniques to build the models that are used to estimate the visual quality level of any image. For the machine learning training phase, two main processes are proposed: The first proposed process consists of a direct learning using all the selected features in only one training phase, named direct learning blind visual quality assessment DLBQA. The second process is an indirect learning and consists of two training phases, named indirect learning blind visual quality assessment ILBQA. This second process includes an additional phase of construction of intermediary metrics used for the construction of the prediction model. The produced models are evaluated on many benchmarks image databases as TID2013, LIVE, and LIVE in the wild image quality challenge. The experimental results demonstrate that the proposed models produce the best visual perception quality prediction, compared to the state-of-the-art models. The proposed models have been implemented on an FPGA platform to demonstrate the feasibility of integrating the proposed solution on an image sensor.
Collapse
|
37
|
No-Reference Image Quality Assessment with Convolutional Neural Networks and Decision Fusion. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app12010101] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
Abstract
No-reference image quality assessment (NR-IQA) has always been a difficult research problem because digital images may suffer very diverse types of distortions and their contents are extremely various. Moreover, IQA is also a very hot topic in the research community since the number and role of digital images in everyday life is continuously growing. Recently, a huge amount of effort has been devoted to exploiting convolutional neural networks and other deep learning techniques for no-reference image quality assessment. Since deep learning relies on a massive amount of labeled data, utilizing pretrained networks has become very popular in the literature. In this study, we introduce a novel, deep learning-based NR-IQA architecture that relies on the decision fusion of multiple image quality scores coming from different types of convolutional neural networks. The main idea behind this scheme is that a diverse set of different types of networks is able to better characterize authentic image distortions than a single network. The experimental results show that our method can effectively estimate perceptual image quality on four large IQA benchmark databases containing either authentic or artificial distortions. These results are also confirmed in significance and cross database tests.
Collapse
|
38
|
Linden Y, Iliffe WR, He G, Danaie M, Fischer DX, Eisterer M, Speller SC, Grovenor CRM. Analysing neutron radiation damage in YBa 2 Cu 3 O 7-x high temperature superconductor tapes. J Microsc 2021; 286:3-12. [PMID: 34879153 DOI: 10.1111/jmi.13078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2021] [Revised: 10/27/2021] [Accepted: 11/29/2021] [Indexed: 11/30/2022]
Abstract
Superconducting windings will be necessary in future fusion reactors to generate the strong magnetic fields needed to confine the plasma, and these superconducting materials will inevitably be exposed to neutron damage. It is known that this exposure results in the creation of isolated damage cascades, but the presence of these defects alone is not sufficient to explain the degradation of macroscopic superconducting properties and a quantitative method is needed to assess the subtle lattice damage in between the clusters. We have studied REBCO coated conductors irradiated with neutrons to a cumulative dose of 3.3×1022 n*m-2 that show a degradation of both Tc and Jc values, and use HRTEM analysis to show that this irradiation introduces ∼10 nm amorphous collision cascades. In addition we introduce a new method for the analysis of these images to quantify the degree of lattice disorder in the apparently perfect matrix between these cascades. This method utilises Fast Fourier and Discrete Cosine Transformations of a statistically-relevant number of HRTEM images of pristine, neutron-irradiated, and amorphous samples, and extracts the degree of randomness in terms of entropy values. Our results show that these entropy values in both mid-frequency band FFT and DCT domains correlate with the expected level of lattice damage, with the pristine samples having the lowest and the fully amorphous regions the highest entropy values. Our methodology allows us to quantify 'invisible' lattice damage to and correlate these values to the degradation of superconducting properties, and also has relevance for a wider range of applications in the field of electron microscopy where small changes in lattice perfection need to be measured. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Y Linden
- Department of Materials, University of Oxford, Parks Rd, Oxford, OX1 3PH, UK
| | - W R Iliffe
- Department of Materials, University of Oxford, Parks Rd, Oxford, OX1 3PH, UK
| | - G He
- Department of Materials, University of Oxford, Parks Rd, Oxford, OX1 3PH, UK
| | - M Danaie
- Electron Physical Sciences Imaging Centre (ePSIC), Diamond Light Source, Didcot, UK
| | - D X Fischer
- Plasma Science and Fusion Center, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - M Eisterer
- Atominstitut, TU Wien, Stadionallee2, A-1020, Vienna, Austria
| | - S C Speller
- Department of Materials, University of Oxford, Parks Rd, Oxford, OX1 3PH, UK
| | - C R M Grovenor
- Department of Materials, University of Oxford, Parks Rd, Oxford, OX1 3PH, UK
| |
Collapse
|
39
|
Wu L, Zhang X, Chen H, Wang D, Deng J. VP-NIQE: An opinion-unaware visual perception natural image quality evaluator. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.08.048] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
40
|
Yu Y, Shi S, Wang Y, Lian X, Liu J, Lei F. Learning to Predict Page View on College Official Accounts With Quality-Aware Features. Front Neurosci 2021; 15:766396. [PMID: 34776856 PMCID: PMC8581399 DOI: 10.3389/fnins.2021.766396] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2021] [Accepted: 09/27/2021] [Indexed: 11/13/2022] Open
Abstract
At present, most of departments in colleges have their own official accounts, which have become the primary channel for announcements and news. In the official accounts, the popularity of articles is influenced by many different factors, such as the content of articles, the aesthetics of the layout, and so on. This paper mainly studies how to learn a computational model for predicting page view on college official accounts with quality-aware features extracted from pictures. First, we built a new picture database by collecting 1,000 pictures from the official accounts of nine well-known universities in the city of Beijing. Then, we proposed a new model for predicting page view by using a selective ensemble technology to fuse three sets of quality-aware features that could represent how a picture looks. Experimental results show that the proposed model has achieved competitive performance against state-of-the-art relevant models on the task for inferring page view from pictures on college official accounts.
Collapse
Affiliation(s)
- Yibing Yu
- The Communist Youth League Committee, Beijing University of Technology, Beijing, China
- School of Economics and Management, Beijing University of Technology, Beijing, China
| | - Shuang Shi
- Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Yifei Wang
- Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Xinkang Lian
- Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Jing Liu
- Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Fei Lei
- Faculty of Information Technology, Beijing University of Technology, Beijing, China
| |
Collapse
|
41
|
Analysis of Benford’s Law for No-Reference Quality Assessment of Natural, Screen-Content, and Synthetic Images. ELECTRONICS 2021. [DOI: 10.3390/electronics10192378] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
With the tremendous growth and usage of digital images, no-reference image quality assessment is becoming increasingly important. This paper presents in-depth analysis of Benford’s law inspired first digit distribution feature vectors for no-reference quality assessment of natural, screen-content, and synthetic images in various viewpoints. Benford’s law makes a prediction for the probability distribution of first digits in natural datasets. It has been applied among others for detecting fraudulent income tax returns, detecting scientific fraud, election forensics, and image forensics. In particular, our analysis is based on first digit distributions in multiple domains (wavelet coefficients, DCT coefficients, singular values, etc.) as feature vectors and the extracted features are mapped onto image quality scores. Extensive experiments have been carried out on seven large image quality benchmark databases. It has been demonstrated that first digit distributions are quality-aware features, and it is possible to reach or outperform the state-of-the-art with them.
Collapse
|
42
|
Yu X, Birkbeck N, Wang Y, Bampis CG, Adsumilli B, Bovik AC. Predicting the Quality of Compressed Videos With Pre-Existing Distortions. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:7511-7526. [PMID: 34460374 DOI: 10.1109/tip.2021.3107213] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Because of the increasing ease of video capture, many millions of consumers create and upload large volumes of User-Generated-Content (UGC) videos to social and streaming media sites over the Internet. UGC videos are commonly captured by naive users having limited skills and imperfect techniques, and tend to be afflicted by mixtures of highly diverse in-capture distortions. These UGC videos are then often uploaded for sharing onto cloud servers, where they are further compressed for storage and transmission. Our paper tackles the highly practical problem of predicting the quality of compressed videos (perhaps during the process of compression, to help guide it), with only (possibly severely) distorted UGC videos as references. To address this problem, we have developed a novel Video Quality Assessment (VQA) framework that we call 1stepVQA (to distinguish it from two-step methods that we discuss). 1stepVQA overcomes limitations of Full-Reference, Reduced-Reference and No-Reference VQA models by exploiting the statistical regularities of both natural videos and distorted videos. We also describe a new dedicated video database, which was created by applying a realistic VMAF-Guided perceptual rate distortion optimization (RDO) criterion to create realistically compressed versions of UGC source videos, which typically have pre-existing distortions. We show that 1stepVQA is able to more accurately predict the quality of compressed videos, given imperfect reference videos, and outperforms other VQA models in this scenario.
Collapse
|
43
|
Abstract
The goal of no-reference image quality assessment (NR-IQA) is to evaluate their perceptual quality of digital images without using the distortion-free, pristine counterparts. NR-IQA is an important part of multimedia signal processing since digital images can undergo a wide variety of distortions during storage, compression, and transmission. In this paper, we propose a novel architecture that extracts deep features from the input image at multiple scales to improve the effectiveness of feature extraction for NR-IQA using convolutional neural networks. Specifically, the proposed method extracts deep activations for local patches at multiple scales and maps them onto perceptual quality scores with the help of trained Gaussian process regressors. Extensive experiments demonstrate that the introduced algorithm performs favorably against the state-of-the-art methods on three large benchmark datasets with authentic distortions (LIVE In the Wild, KonIQ-10k, and SPAQ).
Collapse
|
44
|
Tu Z, Wang Y, Birkbeck N, Adsumilli B, Bovik AC. UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:4449-4464. [PMID: 33856995 DOI: 10.1109/tip.2021.3072221] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Recent years have witnessed an explosion of user-generated content (UGC) videos shared and streamed over the Internet, thanks to the evolution of affordable and reliable consumer capture devices, and the tremendous popularity of social media platforms. Accordingly, there is a great need for accurate video quality assessment (VQA) models for UGC/consumer videos to monitor, control, and optimize this vast content. Blind quality prediction of in-the-wild videos is quite challenging, since the quality degradations of UGC videos are unpredictable, complicated, and often commingled. Here we contribute to advancing the UGC-VQA problem by conducting a comprehensive evaluation of leading no-reference/blind VQA (BVQA) features and models on a fixed evaluation architecture, yielding new empirical insights on both subjective video quality studies and objective VQA model design. By employing a feature selection strategy on top of efficient BVQA models, we are able to extract 60 out of 763 statistical features used in existing methods to create a new fusion-based model, which we dub the VIDeo quality EVALuator (VIDEVAL), that effectively balances the trade-off between VQA performance and efficiency. Our experimental results show that VIDEVAL achieves state-of-the-art performance at considerably lower computational cost than other leading models. Our study protocol also defines a reliable benchmark for the UGC-VQA problem, which we believe will facilitate further research on deep learning-based VQA modeling, as well as perceptually-optimized efficient UGC video processing, transcoding, and streaming. To promote reproducible research and public evaluation, an implementation of VIDEVAL has been made available online: https://github.com/vztu/VIDEVAL.
Collapse
|
45
|
Wu J, Yang W, Li L, Dong W, Shi G, Lin W. Blind image quality prediction with hierarchical feature aggregation. Inf Sci (N Y) 2021. [DOI: 10.1016/j.ins.2020.12.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
46
|
Varga D. No-Reference Image Quality Assessment with Global Statistical Features. J Imaging 2021; 7:29. [PMID: 34460628 PMCID: PMC8321268 DOI: 10.3390/jimaging7020029] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/28/2020] [Revised: 01/22/2021] [Accepted: 02/02/2021] [Indexed: 11/16/2022] Open
Abstract
The perceptual quality of digital images is often deteriorated during storage, compression, and transmission. The most reliable way of assessing image quality is to ask people to provide their opinions on a number of test images. However, this is an expensive and time-consuming process which cannot be applied in real-time systems. In this study, a novel no-reference image quality assessment method is proposed. The introduced method uses a set of novel quality-aware features which globally characterizes the statistics of a given test image, such as extended local fractal dimension distribution feature, extended first digit distribution features using different domains, Bilaplacian features, image moments, and a wide variety of perceptual features. Experimental results are demonstrated on five publicly available benchmark image quality assessment databases: CSIQ, MDID, KADID-10k, LIVE In the Wild, and KonIQ-10k.
Collapse
Affiliation(s)
- Domonkos Varga
- Department of Networked Systems and Services, Budapest University of Technology and Economics, 1111 Budapest, Hungary
| |
Collapse
|
47
|
Stępień I, Obuchowicz R, Piórkowski A, Oszust M. Fusion of Deep Convolutional Neural Networks for No-Reference Magnetic Resonance Image Quality Assessment. SENSORS 2021; 21:s21041043. [PMID: 33546412 PMCID: PMC7913522 DOI: 10.3390/s21041043] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/23/2020] [Revised: 01/28/2021] [Accepted: 01/29/2021] [Indexed: 12/02/2022]
Abstract
The quality of magnetic resonance images may influence the diagnosis and subsequent treatment. Therefore, in this paper, a novel no-reference (NR) magnetic resonance image quality assessment (MRIQA) method is proposed. In the approach, deep convolutional neural network architectures are fused and jointly trained to better capture the characteristics of MR images. Then, to improve the quality prediction performance, the support vector machine regression (SVR) technique is employed on the features generated by fused networks. In the paper, several promising network architectures are introduced, investigated, and experimentally compared with state-of-the-art NR-IQA methods on two representative MRIQA benchmark datasets. One of the datasets is introduced in this work. As the experimental validation reveals, the proposed fusion of networks outperforms related approaches in terms of correlation with subjective opinions of a large number of experienced radiologists.
Collapse
Affiliation(s)
- Igor Stępień
- Doctoral School of Engineering and Technical Sciences at the Rzeszow University of Technology, al. Powstancow Warszawy 12, 35-959 Rzeszow, Poland;
| | - Rafał Obuchowicz
- Department of Diagnostic Imaging, Jagiellonian University Medical College, 19 Kopernika Street, 31-501 Cracow, Poland;
| | - Adam Piórkowski
- Department of Biocybernetics and Biomedical Engineering, AGH University of Science and Technology, al. Mickiewicza 30, 30-059 Cracow, Poland;
| | - Mariusz Oszust
- Department of Computer and Control Engineering, Rzeszow University of Technology, W. Pola 2, 35-959 Rzeszow, Poland
- Correspondence:
| |
Collapse
|
48
|
Shen L, Chen X, Pan Z, Fan K, Li F, Lei J. No-reference stereoscopic image quality assessment based on global and local content characteristics. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2020.10.024] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
49
|
Jiang H, Jiang G, Yu M, Zhang Y, Yang Y, Peng Z, Chen F, Zhang Q. Cubemap-Based Perception-Driven Blind Quality Assessment for 360-degree Images. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:2364-2377. [PMID: 33481711 DOI: 10.1109/tip.2021.3052073] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
image can be represented with different formats, such as the equirectangular projection (ERP) image, viewport images or spherical image, for its different processing procedures and applications. Accordingly, the 360-degree image quality assessment (360-IQA) can be performed on these different formats. However, the performance of 360-IQA with the ERP image is not equivalent with those with the viewport images or spherical image due to the over-sampling and the resulted obvious geometric distortion of ERP image. This imbalance problem brings challenge to ERP image based applications, such as 360-degree image/video compression and assessment. In this paper, we propose a new blind 360-IQA framework to handle this imbalance problem. In the proposed framework, cubemap projection (CMP) with six inter-related faces is used to realize the omnidirectional viewing of 360-degree image. A multi-distortions visual attention quality dataset for 360-degree images is firstly established as the benchmark to analyze the performance of objective 360-IQA methods. Then, the perception-driven blind 360-IQA framework is proposed based on six cubemap faces of CMP for 360-degree image, in which human attention behavior is taken into account to improve the effectiveness of the proposed framework. The cubemap quality feature subset of CMP image is first obtained, and additionally, attention feature matrices and subsets are also calculated to describe the human visual behavior. Experimental results show that the proposed framework achieves superior performances compared with state-of-the-art IQA methods, and the cross dataset validation also verifies the effectiveness of the proposed framework. In addition, the proposed framework can also be combined with new quality feature extraction method to further improve the performance of 360-IQA. All of these demonstrate that the proposed framework is effective in 360-IQA and has a good potential for future applications.
Collapse
|
50
|
Reduced-reference Perceptual Discrepancy Learning for Image Restoration Quality Assessment. ARTIF INTELL 2021. [DOI: 10.1007/978-3-030-93046-2_31] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|