1
|
Chen J, Yan J, Fang Y, Niu L. Webly Supervised Fine-Grained Classification by Integrally Tackling Noises and Subtle Differences. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2025; 34:2641-2653. [PMID: 40279222 DOI: 10.1109/tip.2025.3562740] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/27/2025]
Abstract
Webly-supervised fine-grained visual classification (WSL-FGVC) aims to learn similar sub-classes from cheap web images, which suffers from two major issues: label noises in web images and subtle differences among fine-grained classes. However, existing methods for WSL-FGVC only focus on suppressing noise at image-level, but neglect to mine cues at pixel-level to distinguish the subtle differences among fine-grained classes. In this paper, we propose a bag-level top-down attention framework, which could tackle label noises and mine subtle cues simultaneously and integrally. Specifically, our method first extracts high-level semantic information from a bag of images belonging to the same class, and then uses the bag-level information to mine discriminative regions in various scales of each image. Besides, we propose to derive attention weights from attention maps to weight the bag-level fusion for a robust supervision. We also propose an attention loss on self-bag attention and cross-bag attention to facilitate the learning of valid attention. Extensive experiments on four WSL-FGVC datasets, i.e., Web-Aircraft, Web-Bird, Web-Car, and WebiNat-5089, demonstrate the effectiveness of our method against the state-of-the-art methods.
Collapse
|
2
|
Zhang Z, Liu Z, Ning L, Martin A, Xiong J. Representation of Imprecision in Deep Neural Networks for Image Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:1199-1212. [PMID: 37948150 DOI: 10.1109/tnnls.2023.3329712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/12/2023]
Abstract
Quantification and reduction of uncertainty in deep-learning techniques have received much attention but ignored how to characterize the imprecision caused by such uncertainty. In some tasks, we prefer to obtain an imprecise result rather than being willing or unable to bear the cost of an error. For this purpose, we investigate the representation of imprecision in deep-learning (RIDL) techniques based on the theory of belief functions (TBF). First, the labels of some training images are reconstructed using the learning mechanism of neural networks to characterize the imprecision in the training set. In the process, a label assignment rule is proposed to reassign one or more labels to each training image. Once an image is assigned with multiple labels, it indicates that the image may be in an overlapping region of different categories from the feature perspective or the original label is wrong. Second, those images with multiple labels are rechecked. As a result, the imprecision (multiple labels) caused by the original labeling errors will be corrected, while the imprecision caused by insufficient knowledge is retained. Images with multiple labels are called imprecise ones, and they are considered to belong to meta-categories, the union of some specific categories. Third, the deep network model is retrained based on the reconstructed training set, and the test images are then classified. Finally, some test images that specific categories cannot distinguish will be assigned to meta-categories to characterize the imprecision in the results. Experiments based on some remarkable networks have shown that RIDL can improve accuracy (AC) and reasonably represent imprecision both in the training and testing sets.
Collapse
|
3
|
Zhang Q, Zhu Y, Yang M, Jin G, Zhu Y, Lu Y, Zou Y, Chen Q. An improved sample selection framework for learning with noisy labels. PLoS One 2024; 19:e0309841. [PMID: 39636882 PMCID: PMC11620405 DOI: 10.1371/journal.pone.0309841] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Accepted: 08/20/2024] [Indexed: 12/07/2024] Open
Abstract
Deep neural networks have powerful memory capabilities, yet they frequently suffer from overfitting to noisy labels, leading to a decline in classification and generalization performance. To address this issue, sample selection methods that filter out potentially clean labels have been proposed. However, there is a significant gap in size between the filtered, possibly clean subset and the unlabeled subset, which becomes particularly pronounced at high-noise rates. Consequently, this results in underutilizing label-free samples in sample selection methods, leaving room for performance improvement. This study introduces an enhanced sample selection framework with an oversampling strategy (SOS) to overcome this limitation. This framework leverages the valuable information contained in label-free instances to enhance model performance by combining an SOS with state-of-the-art sample selection methods. We validate the effectiveness of SOS through extensive experiments conducted on both synthetic noisy datasets and real-world datasets such as CIFAR, WebVision, and Clothing1M. The source code for SOS will be made available at https://github.com/LanXiaoPang613/SOS.
Collapse
Affiliation(s)
- Qian Zhang
- School of Information Technology, Jiangsu Open University, Nanjing, Jiangsu, China
| | - Yi Zhu
- School of Information Technology, Jiangsu Open University, Nanjing, Jiangsu, China
| | - Ming Yang
- School of Computer and Electronic Information, Nanjing Normal University, Nanjing, Jiangsu, China
| | - Ge Jin
- School of Information Technology, Jiangsu Open University, Nanjing, Jiangsu, China
| | - Yingwen Zhu
- School of Information Technology, Jiangsu Open University, Nanjing, Jiangsu, China
| | - Yanjun Lu
- School of Information Technology, Jiangsu Open University, Nanjing, Jiangsu, China
| | - Yu Zou
- School of Information Technology, Jiangsu Open University, Nanjing, Jiangsu, China
- School of Artificial Intelligence (School of Future Technology), Nanjing University of Information Science & Technology, Nanjing, Jiangsu, China
| | - Qiu Chen
- Department of Electrical Engineering and Electronics, Graduate School of Engineering, Kogakuin University, Tokyo, Japan
| |
Collapse
|
4
|
Wang J, Xia X, Lan L, Wu X, Yu J, Yang W, Han B, Liu T. Tackling Noisy Labels With Network Parameter Additive Decomposition. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2024; 46:6341-6354. [PMID: 38546996 DOI: 10.1109/tpami.2024.3382138] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Given data with noisy labels, over-parameterized deep networks suffer overfitting mislabeled data, resulting in poor generalization. The memorization effect of deep networks shows that although the networks have the ability to memorize all noisy data, they would first memorize clean training data, and then gradually memorize mislabeled training data. A simple and effective method that exploits the memorization effect to combat noisy labels is early stopping. However, early stopping cannot distinguish the memorization of clean data and mislabeled data, resulting in the network still inevitably overfitting mislabeled data in the early training stage. In this paper, to decouple the memorization of clean data and mislabeled data, and further reduce the side effect of mislabeled data, we perform additive decomposition on network parameters. Namely, all parameters are additively decomposed into two groups, i.e., parameters w are decomposed as w=σ+γ. Afterward, the parameters σ are considered to memorize clean data, while the parameters γ are considered to memorize mislabeled data. Benefiting from the memorization effect, the updates of the parameters σ are encouraged to fully memorize clean data in early training, and then discouraged with the increase of training epochs to reduce interference of mislabeled data. The updates of the parameters γ are the opposite. In testing, only the parameters σ are employed to enhance generalization. Extensive experiments on both simulated and real-world benchmarks confirm the superior performance of our method.
Collapse
|
5
|
Kim H, Lim S, Park M, Kim K, Kang SH, Lee Y. Optimization of Fast Non-Local Means Noise Reduction Algorithm Parameter in Computed Tomographic Phantom Images Using 3D Printing Technology. Diagnostics (Basel) 2024; 14:1589. [PMID: 39125465 PMCID: PMC11312005 DOI: 10.3390/diagnostics14151589] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2024] [Revised: 07/09/2024] [Accepted: 07/17/2024] [Indexed: 08/12/2024] Open
Abstract
Noise in computed tomography (CT) is inevitably generated, which lowers the accuracy of disease diagnosis. The non-local means approach, a software technique for reducing noise, is widely used in medical imaging. In this study, we propose a noise reduction algorithm based on fast non-local means (FNLMs) and apply it to CT images of a phantom created using 3D printing technology. The self-produced phantom was manufactured using filaments with similar density to human brain tissues. To quantitatively evaluate image quality, the contrast-to-noise ratio (CNR), coefficient of variation (COV), and normalized noise power spectrum (NNPS) were calculated. The results demonstrate that the optimized smoothing factors of FNLMs are 0.08, 0.16, 0.22, 0.25, and 0.32 at 0.001, 0.005, 0.01, 0.05, and 0.1 of noise intensities, respectively. In addition, we compared the optimized FNLMs with noisy, local filters and total variation algorithms. As a result, FNLMs showed superior performance compared to various denoising techniques. Particularly, comparing the optimized FNLMs to the noisy images, the CNR improved by 6.53 to 16.34 times, COV improved by 6.55 to 18.28 times, and the NNPS improved by 10-2 mm2 on average. In conclusion, our approach shows significant potential in enhancing CT image quality with anthropomorphic phantoms, thus addressing the noise issue and improving diagnostic accuracy.
Collapse
Affiliation(s)
- Hajin Kim
- Department of Health Science, General Graduate School of Gachon University, 191, Hambakmoe-ro, Yeonsu-gu, Incheon 21936, Republic of Korea; (H.K.); (S.L.); (M.P.)
| | - Sewon Lim
- Department of Health Science, General Graduate School of Gachon University, 191, Hambakmoe-ro, Yeonsu-gu, Incheon 21936, Republic of Korea; (H.K.); (S.L.); (M.P.)
| | - Minji Park
- Department of Health Science, General Graduate School of Gachon University, 191, Hambakmoe-ro, Yeonsu-gu, Incheon 21936, Republic of Korea; (H.K.); (S.L.); (M.P.)
| | - Kyuseok Kim
- Department of Biomedical Engineering, Eulji University, 553, Sanseong-daero, Sujeong-gu, Seongnam-si 13135, Republic of Korea;
| | - Seong-Hyeon Kang
- Department of Biomedical Engineering, Eulji University, 553, Sanseong-daero, Sujeong-gu, Seongnam-si 13135, Republic of Korea;
| | - Youngjin Lee
- Department of Radiological Science, Gachon University, 191, Hambakmoe-ro, Yeonsu-gu, Incheon 21936, Republic of Korea
| |
Collapse
|
6
|
Zhang S, Li JQ, Fujita H, Li YW, Wang DB, Zhu TT, Zhang ML, Liu CY. Student Loss: Towards the Probability Assumption in Inaccurate Supervision. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2024; 46:4460-4475. [PMID: 38261485 DOI: 10.1109/tpami.2024.3357518] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/25/2024]
Abstract
Noisy labels are often encountered in datasets, but learning with them is challenging. Although natural discrepancies between clean and mislabeled samples in a noisy category exist, most techniques in this field still gather them indiscriminately, which leads to their performances being partially robust. In this paper, we reveal both empirically and theoretically that the learning robustness can be improved by assuming deep features with the same labels follow a student distribution, resulting in a more intuitive method called student loss. By embedding the student distribution and exploiting the sharpness of its curve, our method is naturally data-selective and can offer extra strength to resist mislabeled samples. This ability makes clean samples aggregate tightly in the center, while mislabeled samples scatter, even if they share the same label. Additionally, we employ the metric learning strategy and develop a large-margin student (LT) loss for better capability. It should be noted that our approach is the first work that adopts the prior probability assumption in feature representation to decrease the contributions of mislabeled samples. This strategy can enhance various losses to join the student loss family, even if they have been robust losses. Experiments demonstrate that our approach is more effective in inaccurate supervision. Enhanced LT losses significantly outperform various state-of-the-art methods in most cases. Even huge improvements of over 50% can be obtained under some conditions.
Collapse
|
7
|
Wang H, He J, Cui H, Yuan B, Xia Y. Robust Stochastic Neural Ensemble Learning With Noisy Labels for Thoracic Disease Classification. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2180-2190. [PMID: 38265913 DOI: 10.1109/tmi.2024.3357986] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/26/2024]
Abstract
Chest radiography is the most common radiology examination for thoracic disease diagnosis, such as pneumonia. A tremendous number of chest X-rays prompt data-driven deep learning models in constructing computer-aided diagnosis systems for thoracic diseases. However, in realistic radiology practice, a deep learning-based model often suffers from performance degradation when trained on data with noisy labels possibly caused by different types of annotation biases. To this end, we present a novel stochastic neural ensemble learning (SNEL) framework for robust thoracic disease diagnosis using chest X-rays. The core idea of our method is to learn from noisy labels by constructing model ensembles and designing noise-robust loss functions. Specifically, we propose a fast neural ensemble method that collects parameters simultaneously across model instances and along optimization trajectories. Moreover, we propose a loss function that both optimizes a robust measure and characterizes a diversity measure of ensembles. We evaluated our proposed SNEL method on three publicly available hospital-scale chest X-ray datasets. The experimental results indicate that our method outperforms competing methods and demonstrate the effectiveness and robustness of our method in learning from noisy labels. Our code is available at https://github.com/hywang01/SNEL.
Collapse
|
8
|
Song H, Kim M, Park D, Shin Y, Lee JG. Learning From Noisy Labels With Deep Neural Networks: A Survey. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:8135-8153. [PMID: 35254993 DOI: 10.1109/tnnls.2022.3152527] [Citation(s) in RCA: 84] [Impact Index Per Article: 42.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Deep learning has achieved remarkable success in numerous domains with help from large amounts of big data. However, the quality of data labels is a concern because of the lack of high-quality labels in many real-world scenarios. As noisy labels severely degrade the generalization performance of deep neural networks, learning from noisy labels (robust training) is becoming an important task in modern deep learning applications. In this survey, we first describe the problem of learning with label noise from a supervised learning perspective. Next, we provide a comprehensive review of 62 state-of-the-art robust training methods, all of which are categorized into five groups according to their methodological difference, followed by a systematic comparison of six properties used to evaluate their superiority. Subsequently, we perform an in-depth analysis of noise rate estimation and summarize the typically used evaluation methodology, including public noisy datasets and evaluation metrics. Finally, we present several promising research directions that can serve as a guideline for future studies.
Collapse
|
9
|
de Vos BD, Jansen GE, Išgum I. Stochastic co-teaching for training neural networks with unknown levels of label noise. Sci Rep 2023; 13:16875. [PMID: 37803027 PMCID: PMC10558560 DOI: 10.1038/s41598-023-43864-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2023] [Accepted: 09/29/2023] [Indexed: 10/08/2023] Open
Abstract
Label noise hampers supervised training of neural networks. However, data without label noise is often infeasible to attain, especially for medical tasks. Attaining high-quality medical labels would require a pool of experts and their consensus reading, which would be extremely costly. Several methods have been proposed to mitigate the adverse effects of label noise during training. State-of-the-art methods use multiple networks that exploit different decision boundaries to identify label noise. Among the best performing methods is co-teaching. However, co-teaching comes with the requirement of knowing label noise a priori. Hence, we propose a co-teaching method that does not require any prior knowledge about the level of label noise. We introduce stochasticity to select or reject training instances. We have extensively evaluated the method on synthetic experiments with extreme label noise levels and applied it to real-world medical problems of ECG classification and cardiac MRI segmentation. Results show that the approach is robust to its hyperparameter choice and applies to various classification tasks with unknown levels of label noise.
Collapse
Affiliation(s)
- Bob D de Vos
- Department of Biomedical Engineering and Physics, Amsterdam University Medical Center, University of Amsterdam, Amsterdam, the Netherlands
| | - Gino E Jansen
- Department of Biomedical Engineering and Physics, Amsterdam University Medical Center, University of Amsterdam, Amsterdam, the Netherlands.
- Informatics Institute, University of Amsterdam, Amsterdam, the Netherlands.
| | - Ivana Išgum
- Department of Biomedical Engineering and Physics, Amsterdam University Medical Center, University of Amsterdam, Amsterdam, the Netherlands
- Department of Radiology and Nuclear Medicine, Amsterdam University Medical Center, University of Amsterdam, Amsterdam, the Netherlands
- Informatics Institute, University of Amsterdam, Amsterdam, the Netherlands
| |
Collapse
|
10
|
Li W, Yang D, Ma C, Liu L. Identifying novel disease categories through divergence optimization: An approach to prevent misdiagnosis in medical imaging. Comput Biol Med 2023; 165:107403. [PMID: 37688992 DOI: 10.1016/j.compbiomed.2023.107403] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2023] [Revised: 08/09/2023] [Accepted: 08/26/2023] [Indexed: 09/11/2023]
Abstract
Given the significant changes in human lifestyle, the incidence of colon cancer has rapidly increased. The diagnostic process can often be complicated due to symptom similarities between colon cancer and other colon-related diseases. In an effort to minimize misdiagnosis, deep learning-based approaches for colon cancer diagnosis have notably progressed within the field of clinical medicine, offering more precise detection and improved patient outcomes. Despite these advancements, practical application of these techniques continues to encounter two major challenges: 1) due to the need for expert annotation, only a limited number of labels are utilized for diagnosis; and 2) the existence of diverse disease types can lead to misdiagnosis when the model encounters unfamiliar disease categories. To overcome these hurdles, we present a method incorporating Universal Domain Adaptation (UniDA). By optimizing the divergence of samples in the source domain, our method detects noise. Furthermore, to identify categories that are not present in the source domain, we optimize the divergence of unlabeled samples in the target domain. Experimental validation on two gastrointestinal datasets demonstrates that our method surpasses current state-of-the-art domain adaptation techniques in identifying unknown disease classes. It is worth noting that our proposed method is the first work of medical image diagnosis aimed at the identification of unknown categories of diseases.
Collapse
Affiliation(s)
- Wencai Li
- Department of General Surgery, The Second Affiliated Hospital of Shanghai University (Wenzhou Central Hospital), Wenzhou, Zhejiang, 325000, China.
| | - Daqing Yang
- Department of General Surgery, The Second Affiliated Hospital of Shanghai University (Wenzhou Central Hospital), Wenzhou, Zhejiang, 325000, China.
| | - Chao Ma
- School of Digital Media, Shenzhen Institute of Information Technology, Shenzhen, 518172, China.
| | - Lei Liu
- College of Computer Science, Sichuan University, Chengdu, Sichuan, 610065, China.
| |
Collapse
|
11
|
Sumbul G, Demir B. Generative Reasoning Integrated Label Noise Robust Deep Image Representation Learning. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:4529-4542. [PMID: 37440393 DOI: 10.1109/tip.2023.3293776] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/15/2023]
Abstract
The development of deep learning based image representation learning (IRL) methods has attracted great attention for various image understanding problems. Most of these methods require the availability of a set of high quantity and quality of annotated training images, which can be time-consuming, complex and costly to gather. To reduce labeling costs, crowdsourced data, automatic labeling procedures or citizen science projects can be considered. However, such approaches increase the risk of including label noise in training data. It may result in overfitting on noisy labels when discriminative reasoning is employed as in most of the existing methods. This leads to sub-optimal learning procedures, and thus inaccurate characterization of images. To address this issue, in this paper, we introduce a generative reasoning integrated label noise robust deep representation learning (GRID) approach. The proposed GRID approach aims to model the complementary characteristics of discriminative and generative reasoning for IRL under noisy labels. To this end, we first integrate generative reasoning into discriminative reasoning through a supervised variational autoencoder. This allows the proposed GRID approach to automatically detect training samples with noisy labels. Then, through our label noise robust hybrid representation learning strategy, GRID adjusts the whole learning procedure for IRL of these samples through generative reasoning and that of the other samples through discriminative reasoning. Our approach learns discriminative image representations while preventing interference of noisy labels during training independently from the IRL method being selected. Thus, unlike the existing label noise robust methods, GRID does not depend on the type of annotation, label noise, neural network architecture, loss function or learning task, and thus can be directly utilized for various image understanding problems. Experimental results show the effectiveness of the proposed GRID approach compared to the state-of-the-art methods. The code of the proposed approach is publicly available at https://github.com/gencersumbul/GRID.
Collapse
|
12
|
Yao J, Han B, Zhou Z, Zhang Y, Tsang IW. Latent Class-Conditional Noise Model. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:9964-9980. [PMID: 37027688 DOI: 10.1109/tpami.2023.3247629] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Learning with noisy labels has become imperative in the Big Data era, which saves expensive human labors on accurate annotations. Previous noise-transition-based methods have achieved theoretically-grounded performance under the Class-Conditional Noise model (CCN). However, these approaches builds upon an ideal but impractical anchor set available to pre-estimate the noise transition. Even though subsequent works adapt the estimation as a neural layer, the ill-posed stochastic learning of its parameters in back-propagation easily falls into undesired local minimums. We solve this problem by introducing a Latent Class-Conditional Noise model (LCCN) to parameterize the noise transition under a Bayesian framework. By projecting the noise transition into the Dirichlet space, the learning is constrained on a simplex characterized by the complete dataset, instead of some ad-hoc parametric space wrapped by the neural layer. We then deduce a dynamic label regression method for LCCN, whose Gibbs sampler allows us efficiently infer the latent true labels to train the classifier and to model the noise. Our approach safeguards the stable update of the noise transition, which avoids previous arbitrarily tuning from a mini-batch of samples. We further generalize LCCN to different counterparts compatible with open-set noisy labels, semi-supervised learning as well as cross-model training. A range of experiments demonstrate the advantages of LCCN and its variants over the current state-of-the-art methods. The code is available at here.
Collapse
|
13
|
Chen H, Tan W, Li J, Guan P, Wu L, Yan B, Li J, Wang Y. Adaptive Cross Entropy for ultrasmall object detection in Computed Tomography with noisy labels. Comput Biol Med 2022; 147:105763. [DOI: 10.1016/j.compbiomed.2022.105763] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2022] [Revised: 06/09/2022] [Accepted: 06/18/2022] [Indexed: 11/15/2022]
|
14
|
Xu D, Chen R. Meta-Learning for Decoding Neural Activity Data With Noisy Labels. Front Comput Neurosci 2022; 16:913617. [PMID: 35874318 PMCID: PMC9296819 DOI: 10.3389/fncom.2022.913617] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2022] [Accepted: 06/07/2022] [Indexed: 11/26/2022] Open
Abstract
In neural decoding, a behavioral variable is often generated by manual annotation and the annotated labels could contain extensive label noise, leading to poor model generalizability. Tackling the label noise problem in neural decoding can improve model generalizability and robustness. We use a deep neural network based sample reweighting method to tackle this problem. The proposed method reweights training samples by using a small and clean validation dataset to guide learning. We evaluated the sample reweighting method on simulated neural activity data and calcium imaging data of anterior lateral motor cortex. For the simulated data, the proposed method can accurately predict the behavioral variable even in the scenario that 36 percent of samples in the training dataset are mislabeled. For the anterior lateral motor cortex study, the proposed method can predict trial types with F1 score of around 0.85 even 48 percent of training samples are mislabeled.
Collapse
Affiliation(s)
| | - Rong Chen
- Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland School of Medicine, Baltimore, MD, United States
| |
Collapse
|
15
|
Gao M, Feng X, Geng M, Jiang Z, Zhu L, Meng X, Zhou C, Ren Q, Lu Y. Bayesian statistics-guided label refurbishment mechanism: Mitigating label noise in medical image classification. Med Phys 2022; 49:5899-5913. [PMID: 35678232 DOI: 10.1002/mp.15799] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Revised: 03/26/2022] [Accepted: 05/31/2022] [Indexed: 11/12/2022] Open
Abstract
PURPOSE Deep neural networks (DNNs) have been widely applied in medical image classification, benefiting from its powerful mapping capability among medical images. However, these existing deep learning-based methods depend on an enormous amount of carefully labeled images. Meanwhile, noise is inevitably introduced in the labeling process, degrading the performance of models. Hence, it is significant to devise robust training strategies to mitigate label noise in the medical image classification tasks. METHODS In this work, we propose a novel Bayesian statistics-guided label refurbishment mechanism (BLRM) for DNNs to prevent overfitting noisy images. BLRM utilizes maximum a posteriori probability in the Bayesian statistics and the exponentially time-weighted technique to selectively correct the labels of noisy images. The training images are purified gradually with the training epochs when BLRM is activated, further improving classification performance. RESULTS Comprehensive experiments on both synthetic noisy images (public OCT & Messidor datasets) and real-world noisy images (ANIMAL-10N) demonstrate that BLRM refurbishes the noisy labels selectively, curbing the adverse effects of noisy data. Also, the anti-noise BLRMs integrated with DNNs are effective at different noise ratio and are independent of backbone DNN architectures. In addition, BLRM is superior to state-of-the-art comparative methods of anti-noise. CONCLUSIONS These investigations indicate that the proposed BLRM is well capable of mitigating label noise in medical image classification tasks.
Collapse
Affiliation(s)
- Mengdi Gao
- Department of Biomedical Engineering, College of Future Technology, Peking University, Beijing, China.,Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China.,Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China.,Shenzhen Bay Laboratory 5F, Institute of Biomedical Engineering, Shenzhen, China
| | - Ximeng Feng
- Department of Biomedical Engineering, College of Future Technology, Peking University, Beijing, China.,Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China.,Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China.,Shenzhen Bay Laboratory 5F, Institute of Biomedical Engineering, Shenzhen, China
| | - Mufeng Geng
- Department of Biomedical Engineering, College of Future Technology, Peking University, Beijing, China.,Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China.,Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China.,Shenzhen Bay Laboratory 5F, Institute of Biomedical Engineering, Shenzhen, China
| | - Zhe Jiang
- Department of Biomedical Engineering, College of Future Technology, Peking University, Beijing, China.,Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China.,Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China.,Shenzhen Bay Laboratory 5F, Institute of Biomedical Engineering, Shenzhen, China
| | - Lei Zhu
- Department of Biomedical Engineering, College of Future Technology, Peking University, Beijing, China.,Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China.,Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China.,Shenzhen Bay Laboratory 5F, Institute of Biomedical Engineering, Shenzhen, China
| | - Xiangxi Meng
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Key Laboratory for Research and Evaluation of Radiopharmaceuticals (National Medical Products Administration), Department of Nuclear Medicine, Beijing Cancer Hospital & Institute, Beijing, China
| | - Chuanqing Zhou
- Shenzhen Bay Laboratory 5F, Institute of Biomedical Engineering, Shenzhen, China
| | - Qiushi Ren
- Department of Biomedical Engineering, College of Future Technology, Peking University, Beijing, China.,Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China.,Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China.,Shenzhen Bay Laboratory 5F, Institute of Biomedical Engineering, Shenzhen, China
| | - Yanye Lu
- Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China.,Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
| |
Collapse
|
16
|
Xu Q, Yang Z, Jiang Y, Cao X, Yao Y, Huang Q. Not All Samples are Trustworthy: Towards Deep Robust SVP Prediction. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:3154-3169. [PMID: 33373295 DOI: 10.1109/tpami.2020.3047817] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
In this paper, we study the problem of estimating subjective visual properties (SVP) for images, which is an emerging task in Computer Vision. Generally speaking, collecting SVP datasets involves a crowdsourcing process where annotations are obtained from a wide range of online users. Since the process is done without quality control, SVP datasets are known to suffer from noise. This leads to the issue that not all samples are trustworthy. Facing this problem, we need to develop robust models for learning SVP from noisy crowdsourced annotations. In this paper, we construct two general robust learning frameworks for this application. Specifically, in the first framework, we propose a probabilistic framework to explicitly model the sparse unreliable patterns that exist in the dataset. It is noteworthy that we then provide an alternative framework that could reformulate the sparse unreliable patterns as a "contraction" operation over the original loss function. The latter framework leverages not only efficient end-to-end training but also rigorous theoretical analyses. To apply these frameworks, we further provide two models as implementations of the frameworks, where the sparse noise parameters could be interpreted with the HodgeRank theory. Finally, extensive theoretical and empirical studies show the effectiveness of our proposed framework.
Collapse
|
17
|
Giry Fouquet E, Fauvel M, Mallet C. Fast estimation for robust supervised classification with mixture models. Pattern Recognit Lett 2021. [DOI: 10.1016/j.patrec.2021.10.020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
18
|
Wu Y, Liu H, Li R, Sun S, Weile J, Roth FP. Improved pathogenicity prediction for rare human missense variants. Am J Hum Genet 2021; 108:1891-1906. [PMID: 34551312 PMCID: PMC8546039 DOI: 10.1016/j.ajhg.2021.08.012] [Citation(s) in RCA: 71] [Impact Index Per Article: 17.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2021] [Accepted: 08/18/2021] [Indexed: 01/01/2023] Open
Abstract
The success of personalized genomic medicine depends on our ability to assess the pathogenicity of rare human variants, including the important class of missense variation. There are many challenges in training accurate computational systems, e.g., in finding the balance between quantity, quality, and bias in the variant sets used as training examples and avoiding predictive features that can accentuate the effects of bias. Here, we describe VARITY, which judiciously exploits a larger reservoir of training examples with uncertain accuracy and representativity. To limit circularity and bias, VARITY excludes features informed by variant annotation and protein identity. To provide a rationale for each prediction, we quantified the contribution of features and feature combinations to the pathogenicity inference of each variant. VARITY outperformed all previous computational methods evaluated, identifying at least 10% more pathogenic variants at thresholds achieving high (90% precision) stringency.
Collapse
|
19
|
How to handle noisy labels for robust learning from uncertainty. Neural Netw 2021; 143:209-217. [PMID: 34157645 DOI: 10.1016/j.neunet.2021.06.012] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2020] [Revised: 06/07/2021] [Accepted: 06/07/2021] [Indexed: 11/22/2022]
Abstract
Most deep neural networks (DNNs) are trained with large amounts of noisy labels when they are applied. As DNNs have the high capacity to fit any noisy labels, it is known to be difficult to train DNNs robustly with noisy labels. These noisy labels cause the performance degradation of DNNs due to the memorization effect by over-fitting. Earlier state-of-the-art methods used small loss tricks to efficiently resolve the robust training problem with noisy labels. In this paper, relationship between the uncertainties and the clean labels is analyzed. We present novel training method to use not only small loss trick but also labels that are likely to be clean labels selected from uncertainty called "Uncertain Aware Co-Training (UACT)". Our robust learning techniques (UACT) avoid over-fitting the DNNs by extremely noisy labels. By making better use of the uncertainty acquired from the network itself, we achieve good generalization performance. We compare the proposed method to the current state-of-the-art algorithms for noisy versions of MNIST, CIFAR-10, CIFAR-100, T-ImageNet and News to demonstrate its excellence.
Collapse
|
20
|
Algan G, Ulusoy I. Image classification with deep learning in the presence of noisy labels: A survey. Knowl Based Syst 2021. [DOI: 10.1016/j.knosys.2021.106771] [Citation(s) in RCA: 62] [Impact Index Per Article: 15.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
21
|
Zhang CB, Jiang PT, Hou Q, Wei Y, Han Q, Li Z, Cheng MM. Delving Deep Into Label Smoothing. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:5984-5996. [PMID: 34166191 DOI: 10.1109/tip.2021.3089942] [Citation(s) in RCA: 27] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2023]
Abstract
Label smoothing is an effective regularization tool for deep neural networks (DNNs), which generates soft labels by applying a weighted average between the uniform distribution and the hard label. It is often used to reduce the overfitting problem of training DNNs and further improve classification performance. In this paper, we aim to investigate how to generate more reliable soft labels. We present an Online Label Smoothing (OLS) strategy, which generates soft labels based on the statistics of the model prediction for the target category. The proposed OLS constructs a more reasonable probability distribution between the target categories and non-target categories to supervise DNNs. Experiments demonstrate that based on the same classification models, the proposed approach can effectively improve the classification performance on CIFAR-100, ImageNet, and fine-grained datasets. Additionally, the proposed method can significantly improve the robustness of DNN models to noisy labels compared to current label smoothing approaches. The source code is available at our project page: https://mmcheng.net/ols/.
Collapse
|
22
|
Liu S, Thung KH, Lin W, Shen D, Yap PT. Hierarchical Nonlocal Residual Networks for Image Quality Assessment of Pediatric Diffusion MRI With Limited and Noisy Annotations. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3691-3702. [PMID: 32746115 PMCID: PMC7606371 DOI: 10.1109/tmi.2020.3002708] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Fast and automated image quality assessment (IQA) of diffusion MR images is crucial for making timely decisions for rescans. However, learning a model for this task is challenging as the number of annotated data is limited and the annotation labels might not always be correct. As a remedy, we will introduce in this paper an automatic image quality assessment (IQA) method based on hierarchical non-local residual networks for pediatric diffusion MR images. Our IQA is performed in three sequential stages, i.e., 1) slice-wise IQA, where a nonlocal residual network is first pre-trained to annotate each slice with an initial quality rating (i.e., pass/questionable/fail), which is subsequently refined via iterative semi-supervised learning and slice self-training; 2) volume-wise IQA, which agglomerates the features extracted from the slices of a volume, and uses a nonlocal network to annotate the quality rating for each volume via iterative volume self-training; and 3) subject-wise IQA, which ensembles the volumetric IQA results to determine the overall image quality pertaining to a subject. Experimental results demonstrate that our method, trained using only samples of modest size, exhibits great generalizability, and is capable of conducting rapid hierarchical IQA with near-perfect accuracy.
Collapse
|
23
|
Karimi D, Dou H, Warfield SK, Gholipour A. Deep learning with noisy labels: Exploring techniques and remedies in medical image analysis. Med Image Anal 2020; 65:101759. [PMID: 32623277 PMCID: PMC7484266 DOI: 10.1016/j.media.2020.101759] [Citation(s) in RCA: 202] [Impact Index Per Article: 40.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2019] [Revised: 03/15/2020] [Accepted: 06/16/2020] [Indexed: 01/19/2023]
Abstract
Supervised training of deep learning models requires large labeled datasets. There is a growing interest in obtaining such datasets for medical image analysis applications. However, the impact of label noise has not received sufficient attention. Recent studies have shown that label noise can significantly impact the performance of deep learning models in many machine learning and computer vision applications. This is especially concerning for medical applications, where datasets are typically small, labeling requires domain expertise and suffers from high inter- and intra-observer variability, and erroneous predictions may influence decisions that directly impact human health. In this paper, we first review the state-of-the-art in handling label noise in deep learning. Then, we review studies that have dealt with label noise in deep learning for medical image analysis. Our review shows that recent progress on handling label noise in deep learning has gone largely unnoticed by the medical image analysis community. To help achieve a better understanding of the extent of the problem and its potential remedies, we conducted experiments with three medical imaging datasets with different types of label noise, where we investigated several existing strategies and developed new methods to combat the negative effect of label noise. Based on the results of these experiments and our review of the literature, we have made recommendations on methods that can be used to alleviate the effects of different types of label noise on deep models trained for medical image analysis. We hope that this article helps the medical image analysis researchers and developers in choosing and devising new techniques that effectively handle label noise in deep learning.
Collapse
Affiliation(s)
- Davood Karimi
- Department of Radiology, Boston Children's Hospital, Harvard Medical School, Boston, MA, USA.
| | - Haoran Dou
- Department of Radiology, Boston Children's Hospital, Harvard Medical School, Boston, MA, USA
| | - Simon K Warfield
- Department of Radiology, Boston Children's Hospital, Harvard Medical School, Boston, MA, USA
| | - Ali Gholipour
- Department of Radiology, Boston Children's Hospital, Harvard Medical School, Boston, MA, USA
| |
Collapse
|
24
|
Yao Y, Shen F, Xie G, Liu L, Zhu F, Zhang J, Shen HT. Exploiting Web Images for Multi-Output Classification: From Category to Subcategories. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:2348-2360. [PMID: 32012029 DOI: 10.1109/tnnls.2020.2966644] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Studies present that dividing categories into subcategories contributes to better image classification. Existing image subcategorization works relying on expert knowledge and labeled images are both time-consuming and labor-intensive. In this article, we propose to select and subsequently classify images into categories and subcategories. Specifically, we first obtain a list of candidate subcategory labels from untagged corpora. Then, we purify these subcategory labels through calculating the relevance to the target category. To suppress the search error and noisy subcategory label-induced outlier images, we formulate outlier images removing and the optimal classification models learning as a unified problem to jointly learn multiple classifiers, where the classifier for a category is obtained by combining multiple subcategory classifiers. Compared with the existing subcategorization works, our approach eliminates the dependence on expert knowledge and labeled images. Extensive experiments on image categorization and subcategorization demonstrate the superiority of our proposed approach.
Collapse
|