1
|
Wang Y, Liu C, Fan Y, Niu C, Huang W, Pan Y, Li J, Wang Y, Li J. A multi-modal deep learning solution for precise pneumonia diagnosis: the PneumoFusion-Net model. Front Physiol 2025; 16:1512835. [PMID: 40144549 PMCID: PMC11937601 DOI: 10.3389/fphys.2025.1512835] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2024] [Accepted: 02/14/2025] [Indexed: 03/28/2025] Open
Abstract
Background Pneumonia is considered one of the most important causes of morbidity and mortality in the world. Bacterial and viral pneumonia share many similar clinical features, thus making diagnosis a challenging task. Traditional diagnostic method developments mainly rely on radiological imaging and require a certain degree of consulting clinical experience, which can be inefficient and inconsistent. Deep learning for the classification of pneumonia in multiple modalities, especially integrating multiple data, has not been well explored. Methods The study introduce the PneumoFusion-Net, a deep learning-based multimodal framework that incorporates CT images, clinical text, numerical lab test results, and radiology reports for improved diagnosis. In the experiments, a dataset of 10,095 pneumonia CT images was used-including associated clinical data-most of which was used for training and validation while keeping part of it for validation on a held-out test set. Five-fold cross-validation was considered in order to evaluate this model, calculating different metrics including accuracy and F1-Score. Results PneumoFusion-Net, which achieved 98.96% classification accuracy with a 98% F1-score on the held-out test set, is highly effective in distinguishing bacterial from viral types of pneumonia. This has been highly beneficial for diagnosis, reducing misdiagnosis and further improving homogeneity across various data sets from multiple patients. Conclusion PneumoFusion-Net offers an effective and efficient approach to pneumonia classification by integrating diverse data sources, resulting in high diagnostic accuracy. Its potential for clinical integration could significantly reduce the burden of pneumonia diagnosis by providing radiologists and clinicians with a robust, automated diagnostic tool.
Collapse
Affiliation(s)
- Yujie Wang
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
- Deep Vision Agriculture Lab, Sichuan Agricultural University, Ya’an, China
| | - Can Liu
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
| | - Yinghan Fan
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
| | - Chenyue Niu
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
| | - Wanyun Huang
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
| | - Yixuan Pan
- College of Science, Sichuan Agricultural University, Ya’an, China
| | - Jingze Li
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
- Deep Vision Agriculture Lab, Sichuan Agricultural University, Ya’an, China
| | - Yilin Wang
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
| | - Jun Li
- College of Information Engineering, Sichuan Agricultural University, Ya’an, China
- Agricultural Information Engineering Higher Institution Key Laboratory of Sichuan Province, Ya’an, China
- Ya’an Digital Agricultural Engineering Technology Research Center, Ya’an, China
| |
Collapse
|
2
|
Jemal I, Naoussi Sijou WA, Chikhaoui B. Multi-modal recommender system for predicting project manager performance within a competency-based framework. Front Big Data 2024; 7:1295009. [PMID: 38784678 PMCID: PMC11111973 DOI: 10.3389/fdata.2024.1295009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Accepted: 04/16/2024] [Indexed: 05/25/2024] Open
Abstract
The evaluation of performance using competencies within a structured framework holds significant importance across various professional domains, particularly in roles like project manager. Typically, this assessment process, overseen by senior evaluators, involves scoring competencies based on data gathered from interviews, completed forms, and evaluation programs. However, this task is tedious and time-consuming, and requires the expertise of qualified professionals. Moreover, it is compounded by the inconsistent scoring biases introduced by different evaluators. In this paper, we propose a novel approach to automatically predict competency scores, thereby facilitating the assessment of project managers' performance. Initially, we performed data fusion to compile a comprehensive dataset from various sources and modalities, including demographic data, profile-related data, and historical competency assessments. Subsequently, NLP techniques were used to pre-process text data. Finally, recommender systems were explored to predict competency scores. We compared four different recommender system approaches: content-based filtering, demographic filtering, collaborative filtering, and hybrid filtering. Using assessment data collected from 38 project managers, encompassing scores across 67 different competencies, we evaluated the performance of each approach. Notably, the content-based approach yielded promising results, achieving a precision rate of 81.03%. Furthermore, we addressed the challenge of cold-starting, which in our context involves predicting scores for either a new project manager lacking competency data or a newly introduced competency without historical records. Our analysis revealed that demographic filtering achieved an average precision of 54.05% when dealing with new project managers. In contrast, content-based filtering exhibited remarkable performance, achieving a precision of 85.79% in predicting scores for new competencies. These findings underscore the potential of recommender systems in competency assessment, thereby facilitating more effective performance evaluation process.
Collapse
|
3
|
Research on Recurrence Plot Feature Quantization Method Based on Image Texture Analysis. JOURNAL OF ENVIRONMENTAL AND PUBLIC HEALTH 2022; 2022:2495024. [PMID: 35978591 PMCID: PMC9377861 DOI: 10.1155/2022/2495024] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/11/2022] [Revised: 06/15/2022] [Accepted: 07/09/2022] [Indexed: 11/17/2022]
Abstract
The nonlinear time-series analysis method, based on the recurrence plot theory, has received great attention from researchers and has been successfully used in multiple fields. However, traditional recurrence plots that use Heaviside step functions to determine the recursive behavior of a point in the phase space have two problems: (1) Heaviside step functions produce a rigid boundary, resulting in information loss; and (2) the selection of the critical distance, ε, is crucial; if the selection is inappropriate, it will result in a low-dimensional dynamics error, and as of now, there exists no unified method for selecting this parameter. With regard to the problems described above, the novelty of this article lies in the following: (1) when determining the state-phase point recursiveness, a Gaussian function is used to replace the Heaviside function, thereby solving the rigidity and binary value problems of the recursive analysis results caused by the Heaviside step function; and (2) texture analysis is performed on a recurrence plot, new ways of studying complex system dynamics features are proposed, and a system of complex system dynamic-like measurement methods is built.
Collapse
|
4
|
Tasnim H, Dutta S, Turton TL, Rogers DH, Moses ME. Information-Theoretic Exploration of Multivariate Time-Varying Image Databases. Comput Sci Eng 2022. [DOI: 10.1109/mcse.2022.3188291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Affiliation(s)
| | - Soumya Dutta
- Los Alamos National Laboratory, Los Alamos, NM, USA
| | | | | | | |
Collapse
|
5
|
A Systematic Literature Review on Crop Yield Prediction with Deep Learning and Remote Sensing. REMOTE SENSING 2022. [DOI: 10.3390/rs14091990] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Deep learning has emerged as a potential tool for crop yield prediction, allowing the model to automatically extract features and learn from the datasets. Meanwhile, smart farming technology enables the farmers to achieve maximum crop yield by extracting essential parameters of crop growth. This systematic literature review highlights the existing research gaps in a particular area of deep learning methodologies and guides us in analyzing the impact of vegetation indices and environmental factors on crop yield. To achieve the aims of this study, prior studies from 2012 to 2022 from various databases are collected and analyzed. The study focuses on the advantages of using deep learning in crop yield prediction, the suitable remote sensing technology based on the data acquisition requirements, and the various features that influence crop yield prediction. This study finds that Long Short-Term Memory (LSTM) and Convolutional Neural Networks (CNN) are the most widely used deep learning approaches for crop yield prediction. The commonly used remote sensing technology is satellite remote sensing technology—in particular, the use of the Moderate-Resolution Imaging Spectroradiometer (MODIS). Findings show that vegetation indices are the most used feature for crop yield prediction. However, it is also observed that the most used features in the literature do not always work for all the approaches. The main challenges of using deep learning approaches and remote sensing for crop yield prediction are how to improve the working model for better accuracy, the practical implication of the model for providing accurate information about crop yield to agriculturalists, growers, and policymakers, and the issue with the black box property.
Collapse
|
6
|
A Bounded Measure for Estimating the Benefit of Visualization (Part I): Theoretical Discourse and Conceptual Evaluation. ENTROPY 2022; 24:e24020228. [PMID: 35205522 PMCID: PMC8870844 DOI: 10.3390/e24020228] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/30/2021] [Revised: 01/26/2022] [Accepted: 01/27/2022] [Indexed: 12/10/2022]
Abstract
Information theory can be used to analyze the cost–benefit of visualization processes. However, the current measure of benefit contains an unbounded term that is neither easy to estimate nor intuitive to interpret. In this work, we propose to revise the existing cost–benefit measure by replacing the unbounded term with a bounded one. We examine a number of bounded measures that include the Jenson–Shannon divergence, its square root, and a new divergence measure formulated as part of this work. We describe the rationale for proposing a new divergence measure. In the first part of this paper, we focus on the conceptual analysis of the mathematical properties of these candidate measures. We use visualization to support the multi-criteria comparison, narrowing the search down to several options with better mathematical properties. The theoretical discourse and conceptual evaluation in this part provides the basis for further data-driven evaluation based on synthetic and experimental case studies that are reported in the second part of this paper.
Collapse
|
7
|
Biswas A, Dutta S, Lawrence E, Patchett J, Calhoun JC, Ahrens J. Probabilistic Data-Driven Sampling via Multi-Criteria Importance Analysis. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2021; 27:4439-4454. [PMID: 32746272 DOI: 10.1109/tvcg.2020.3006426] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Although supercomputers are becoming increasingly powerful, their components have thus far not scaled proportionately. Compute power is growing enormously and is enabling finely resolved simulations that produce never-before-seen features. However, I/O capabilities lag by orders of magnitude, which means only a fraction of the simulation data can be stored for post hoc analysis. Prespecified plans for saving features and quantities of interest do not work for features that have not been seen before. Data-driven intelligent sampling schemes are needed to detect and save important parts of the simulation while it is running. Here, we propose a novel sampling scheme that reduces the size of the data by orders-of-magnitude while still preserving important regions. The approach we develop selects points with unusual data values and high gradients. We demonstrate that our approach outperforms traditional sampling schemes on a number of tasks.
Collapse
|
8
|
Chen Z, Xu Z, Gui Q, Yang X, Cheng Q, Hou W, Ding M. Self-learning based medical image representation for rigid real-time and multimodal slice-to-volume registration. Inf Sci (N Y) 2020. [DOI: 10.1016/j.ins.2020.06.072] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
9
|
Gao J, Li P, Chen Z, Zhang J. A Survey on Deep Learning for Multimodal Data Fusion. Neural Comput 2020; 32:829-864. [PMID: 32186998 DOI: 10.1162/neco_a_01273] [Citation(s) in RCA: 73] [Impact Index Per Article: 14.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
With the wide deployments of heterogeneous networks, huge amounts of data with characteristics of high volume, high variety, high velocity, and high veracity are generated. These data, referred to multimodal big data, contain abundant intermodality and cross-modality information and pose vast challenges on traditional data fusion methods. In this review, we present some pioneering deep learning models to fuse these multimodal big data. With the increasing exploration of the multimodal big data, there are still some challenges to be addressed. Thus, this review presents a survey on deep learning for multimodal data fusion to provide readers, regardless of their original community, with the fundamentals of multimodal deep learning fusion method and to motivate new multimodal data fusion techniques of deep learning. Specifically, representative architectures that are widely used are summarized as fundamental to the understanding of multimodal deep learning. Then the current pioneering multimodal data fusion deep learning models are summarized. Finally, some challenges and future topics of multimodal data fusion deep learning models are described.
Collapse
Affiliation(s)
- Jing Gao
- School of Software Technology, Dalian University of Technology, Dalian 116620, China, and Key Laboratory for Ubiquitous Network and Service Software of Liaoning Province, Dalian 116620, China
| | - Peng Li
- School of Software Technology, Dalian University of Technology, Dalian 116620, China
| | - Zhikui Chen
- School of Software Technology, Dalian University of Technology, Dalian 116620, China, and Key Laboratory for Ubiquitous Network and Service Software of Liaoning Province, Dalian 116620, China
| | - Jianing Zhang
- School of Software Technology, Dalian University of Technology, Dalian 116620, China
| |
Collapse
|
10
|
Dutta S, Biswas A, Ahrens J. Multivariate Pointwise Information-Driven Data Sampling and Visualization. ENTROPY 2019; 21:e21070699. [PMID: 33267413 PMCID: PMC7515213 DOI: 10.3390/e21070699] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/08/2019] [Revised: 06/25/2019] [Accepted: 07/06/2019] [Indexed: 12/05/2022]
Abstract
With increasing computing capabilities of modern supercomputers, the size of the data generated from the scientific simulations is growing rapidly. As a result, application scientists need effective data summarization techniques that can reduce large-scale multivariate spatiotemporal data sets while preserving the important data properties so that the reduced data can answer domain-specific queries involving multiple variables with sufficient accuracy. While analyzing complex scientific events, domain experts often analyze and visualize two or more variables together to obtain a better understanding of the characteristics of the data features. Therefore, data summarization techniques are required to analyze multi-variable relationships in detail and then perform data reduction such that the important features involving multiple variables are preserved in the reduced data. To achieve this, in this work, we propose a data sub-sampling algorithm for performing statistical data summarization that leverages pointwise information theoretic measures to quantify the statistical association of data points considering multiple variables and generates a sub-sampled data that preserves the statistical association among multi-variables. Using such reduced sampled data, we show that multivariate feature query and analysis can be done effectively. The efficacy of the proposed multivariate association driven sampling algorithm is presented by applying it on several scientific data sets.
Collapse
|
11
|
Information Guided Exploration of Scalar Values and Isocontours in Ensemble Datasets. ENTROPY 2018; 20:e20070540. [PMID: 33265629 PMCID: PMC7513067 DOI: 10.3390/e20070540] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/26/2018] [Revised: 07/16/2018] [Accepted: 07/18/2018] [Indexed: 11/23/2022]
Abstract
Uncertainty of scalar values in an ensemble dataset is often represented by the collection of their corresponding isocontours. Various techniques such as contour-boxplot, contour variability plot, glyphs and probabilistic marching-cubes have been proposed to analyze and visualize ensemble isocontours. All these techniques assume that a scalar value of interest is already known to the user. Not much work has been done in guiding users to select the scalar values for such uncertainty analysis. Moreover, analyzing and visualizing a large collection of ensemble isocontours for a selected scalar value has its own challenges. Interpreting the visualizations of such large collections of isocontours is also a difficult task. In this work, we propose a new information-theoretic approach towards addressing these issues. Using specific information measures that estimate the predictability and surprise of specific scalar values, we evaluate the overall uncertainty associated with all the scalar values in an ensemble system. This helps the scientist to understand the effects of uncertainty on different data features. To understand in finer details the contribution of individual members towards the uncertainty of the ensemble isocontours of a selected scalar value, we propose a conditional entropy based algorithm to quantify the individual contributions. This can help simplify analysis and visualization for systems with more members by identifying the members contributing the most towards overall uncertainty. We demonstrate the efficacy of our method by applying it on real-world datasets from material sciences, weather forecasting and ocean simulation experiments.
Collapse
|
12
|
|
13
|
Multicomponent and Longitudinal Imaging Seen as a Communication Channel—An Application to Stroke. ENTROPY 2017. [DOI: 10.3390/e19050187] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
14
|
Chen M, Golan A. What May Visualization Processes Optimize? IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2016; 22:2619-2632. [PMID: 26731770 DOI: 10.1109/tvcg.2015.2513410] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/07/2023]
Abstract
In this paper, we present an abstract model of visualization and inference processes, and describe an information-theoretic measure for optimizing such processes. In order to obtain such an abstraction, we first examined six classes of workflows in data analysis and visualization, and identified four levels of typical visualization components, namely disseminative, observational, analytical and model-developmental visualization. We noticed a common phenomenon at different levels of visualization, that is, the transformation of data spaces (referred to as alphabets) usually corresponds to the reduction of maximal entropy along a workflow. Based on this observation, we establish an information-theoretic measure of cost-benefit ratio that may be used as a cost function for optimizing a data visualization process. To demonstrate the validity of this measure, we examined a number of successful visualization processes in the literature, and showed that the information-theoretic measure can mathematically explain the advantages of such processes over possible alternatives.
Collapse
|
15
|
Biswas A, Dutta S, Shen HW, Woodring J. An information-aware framework for exploring multivariate data sets. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2013; 19:2683-2692. [PMID: 24051835 DOI: 10.1109/tvcg.2013.133] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
Information theory provides a theoretical framework for measuring information content for an observed variable, and has attracted much attention from visualization researchers for its ability to quantify saliency and similarity among variables. In this paper, we present a new approach towards building an exploration framework based on information theory to guide the users through the multivariate data exploration process. In our framework, we compute the total entropy of the multivariate data set and identify the contribution of individual variables to the total entropy. The variables are classified into groups based on a novel graph model where a node represents a variable and the links encode the mutual information shared between the variables. The variables inside the groups are analyzed for their representativeness and an information based importance is assigned. We exploit specific information metrics to analyze the relationship between the variables and use the metrics to choose isocontours of selected variables. For a chosen group of points, parallel coordinates plots (PCP) are used to show the states of the variables and provide an interface for the user to select values of interest. Experiments with different data sets reveal the effectiveness of our proposed framework in depicting the interesting regions of the data sets taking into account the interaction among the variables.
Collapse
|
16
|
Bramon R, Ruiz M, Bardera A, Boada I, Feixas M, Sbert M. Information Theory-Based Automatic Multimodal Transfer Function Design. IEEE J Biomed Health Inform 2013; 17:870-80. [DOI: 10.1109/jbhi.2013.2263227] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|