1
|
Wang J, Zang J, An Q, Wang H, Zhang Z. A pooling convolution model for multi-classification of ECG and PCG signals. Comput Methods Biomech Biomed Engin 2025; 28:628-641. [PMID: 38193152 DOI: 10.1080/10255842.2023.2299697] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Revised: 11/22/2023] [Accepted: 12/08/2023] [Indexed: 01/10/2024]
Abstract
Electrocardiogram (ECG) and phonocardiogram (PCG) signals are physiological signals generated throughout the cardiac cycle. The application of deep learning techniques to recognize ECG and PCG signals can greatly enhance the efficiency of cardiovascular disease detection. Therefore, we propose a series of straightforward and effective pooling convolutional models for the multi-classification of ECG and PCG signals. Initially, these signals undergo preprocessing. Subsequently, we design various structural blocks, including a stacked block (MCM) comprising convolutional layer and max-pooling layers, along with its variations, as well as a residual block (REC). By adjusting the number of structural blocks, these models can handle ECG and PCG data with different sampling rates. In the final tests, the models utilizing the MCM structural block achieved accuracies of 98.70 and 92.58% on the ECG and PCG fusion datasets, respectively. These accuracies surpass those of all networks utilizing its variations. Moreover, compared to the models employing the REC structural block, the accuracies are improved by 0.02 and 4.30%, respectively. Furthermore, this research has been validated through tests conducted on multiple ECG and PCG datasets, along with comparisons to other published literature. To further validate the generalizability of the model, an additional experiment involving the classification of a synchronized ECG-PCG dataset was conducted. This dataset is divided into seven different levels of fatigue based on the amount of exercise performed by each healthy subject during the testing process. The results indicate that the model using the MCM block also achieved the highest accuracy.
Collapse
Affiliation(s)
- Juliang Wang
- Key Laboratory of Instrumentation Science & Dynamic Measurement of Ministry of Education, North University of China, Taiyuan, China
| | - Junbin Zang
- Key Laboratory of Instrumentation Science & Dynamic Measurement of Ministry of Education, North University of China, Taiyuan, China
| | - Qi An
- Key Laboratory of Instrumentation Science & Dynamic Measurement of Ministry of Education, North University of China, Taiyuan, China
| | - Haoxin Wang
- Key Laboratory of Instrumentation Science & Dynamic Measurement of Ministry of Education, North University of China, Taiyuan, China
| | - Zhidong Zhang
- Key Laboratory of Instrumentation Science & Dynamic Measurement of Ministry of Education, North University of China, Taiyuan, China
| |
Collapse
|
2
|
Li H, Chung H. Prediction of Member Forces of Steel Tubes on the Basis of a Sensor System with the Use of AI. SENSORS (BASEL, SWITZERLAND) 2025; 25:919. [PMID: 39943558 PMCID: PMC11820130 DOI: 10.3390/s25030919] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/19/2024] [Revised: 02/01/2025] [Accepted: 02/02/2025] [Indexed: 02/16/2025]
Abstract
The rapid development of AI (artificial intelligence), sensor technology, high-speed Internet, and cloud computing has demonstrated the potential of data-driven approaches in structural health monitoring (SHM) within the field of structural engineering. Algorithms based on machine learning (ML) models are capable of discerning intricate structural behavioral patterns from real-time data gathered by sensors, thereby offering solutions to engineering quandaries in structural mechanics and SHM. This study presents an innovative approach based on AI and a fiber-reinforced polymer (FRP) double-helix sensor system for the prediction of forces acting on steel tube members in offshore wind turbine support systems; this enables structural health monitoring of the support system. The steel tube as the transitional member and the FRP double helix-sensor system were initially modeled in three dimensions using ABAQUS finite element software. Subsequently, the data obtained from the finite element analysis (FEA) were inputted into a fully connected neural network (FCNN) model, with the objective of establishing a nonlinear mapping relationship between the inputs (strain) and the outputs (reaction force). In the FCNN model, the impact of the number of input variables on the model's predictive performance is examined through cross-comparison of different combinations and positions of the six sets of input variables. And based on an evaluation of engineering costs and the number of strain sensors, a series of potential combinations of variables are identified for further optimization. Furthermore, the potential variable combinations were optimized using a convolutional neural network (CNN) model, resulting in optimal input variable combinations that achieved the accuracy level of more input variable combinations with fewer sensors. This not only improves the prediction performance of the model but also effectively controls the engineering cost. The model performance was evaluated using several metrics, including R2, MSE, MAE, and SMAPE. The results demonstrated that the CNN model exhibited notable advantages in terms of fitting accuracy and computational efficiency when confronted with a limited data set. To provide further support for practical applications, an interactive graphical user interface (GUI)-based sensor-coupled mechanical prediction system for steel tubes was developed. This system enables engineers to predict the member forces of steel tubes in real time, thereby enhancing the efficiency and accuracy of SHM for offshore wind turbine support systems.
Collapse
Affiliation(s)
| | - Heungjin Chung
- Artificial Intelligence, Civil and Environmental Engineering, Jeonju University, Jeonju 55069, Republic of Korea;
| |
Collapse
|
3
|
Wang K, Zhu M, Chen Z, Weng J, Li M, Yiu SM, Ding W, Gu T. A Statistical Physics Perspective: Understanding the Causality Behind Convolutional Neural Network Adversarial Vulnerability. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:2118-2132. [PMID: 38324429 DOI: 10.1109/tnnls.2024.3359269] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/09/2024]
Abstract
The adversarial vulnerability of convolutional neural networks (CNNs) refers to the performance degradation of CNNs under adversarial attacks, leading to incorrect decisions. However, the causes of adversarial vulnerability in CNNs remain unknown. To address this issue, we propose a unique cross-scale analytical approach from a statistical physics perspective. It reveals that the huge amount of nonlinear effects inherent in CNNs is the fundamental cause for the formation and evolution of system vulnerability. Vulnerability is spontaneously formed on the macroscopic level after the symmetry of the system is broken through the nonlinear interaction between microscopic state order parameters. We develop a cascade failure algorithm, visualizing how micro perturbations on neurons' activation can cascade and influence macro decision paths. Our empirical results demonstrate the interplay between microlevel activation maps and macrolevel decision-making and provide a statistical physics perspective to understand the causality behind CNN vulnerability. Our work will help subsequent research to improve the adversarial robustness of CNNs.
Collapse
|
4
|
Hu S, Jiang O, Riedmiller J, Bearfield CX. Motion-Based Visual Encoding Can Improve Performance on Perceptual Tasks with Dynamic Time Series. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2025; 31:163-173. [PMID: 39250377 DOI: 10.1109/tvcg.2024.3456405] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/11/2024]
Abstract
Dynamic data visualizations can convey large amounts of information over time, such as using motion to depict changes in data values for multiple entities. Such dynamic displays put a demand on our visual processing capacities, yet our perception of motion is limited. Several techniques have been shown to improve the processing of dynamic displays. Staging the animation to sequentially show steps in a transition and tracing object movement by displaying trajectory histories can improve processing by reducing the cognitive load. In this paper, We examine the effectiveness of staging and tracing in dynamic displays. We showed participants animated line charts depicting the movements of lines and asked them to identify the line with the highest mean and variance. We manipulated the animation to display the lines with or without staging, tracing and history, and compared the results to a static chart as a control. Results showed that tracing and staging are preferred by participants, and improve their performance in mean and variance tasks respectively. They also preferred display time 3 times shorter when staging is used. Also, encoding animation speed with mean and variance in congruent tasks is associated with higher accuracy. These findings help inform real-world best practices for building dynamic displays. The supplementary materials can be found at https://osf.io/8c95v/.
Collapse
|
5
|
Li G, Wang J, Wang Y, Shan G, Zhao Y. An In-Situ Visual Analytics Framework for Deep Neural Networks. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2024; 30:6770-6786. [PMID: 38051629 DOI: 10.1109/tvcg.2023.3339585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
The past decade has witnessed the superior power of deep neural networks (DNNs) in applications across various domains. However, training a high-quality DNN remains a non-trivial task due to its massive number of parameters. Visualization has shown great potential in addressing this situation, as evidenced by numerous recent visualization works that aid in DNN training and interpretation. These works commonly employ a strategy of logging training-related data and conducting post-hoc analysis. Based on the results of offline analysis, the model can be further trained or fine-tuned. This strategy, however, does not cope with the increasing complexity of DNNs, because (1) the time-series data collected over the training are usually too large to be stored entirely; (2) the huge I/O overhead significantly impacts the training efficiency; (3) post-hoc analysis does not allow rapid human-interventions (e.g., stop training with improper hyper-parameter settings to save computational resources). To address these challenges, we propose an in-situ visualization and analysis framework for the training of DNNs. Specifically, we employ feature extraction algorithms to reduce the size of training-related data in-situ and use the reduced data for real-time visual analytics. The states of model training are disclosed to model designers in real-time, enabling human interventions on demand to steer the training. Through concrete case studies, we demonstrate how our in-situ framework helps deep learning experts optimize DNNs and improve their analysis efficiency.
Collapse
|
6
|
Coscia A, Endert A. KnowledgeVIS: Interpreting Language Models by Comparing Fill-in-the-Blank Prompts. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2024; 30:6520-6532. [PMID: 38145514 DOI: 10.1109/tvcg.2023.3346713] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/27/2023]
Abstract
Recent growth in the popularity of large language models has led to their increased usage for summarizing, predicting, and generating text, making it vital to help researchers and engineers understand how and why they work. We present KnowledgeVIS, a human-in-the-loop visual analytics system for interpreting language models using fill-in-the-blank sentences as prompts. By comparing predictions between sentences, KnowledgeVIS reveals learned associations that intuitively connect what language models learn during training to natural language tasks downstream, helping users create and test multiple prompt variations, analyze predicted words using a novel semantic clustering technique, and discover insights using interactive visualizations. Collectively, these visualizations help users identify the likelihood and uniqueness of individual predictions, compare sets of predictions between prompts, and summarize patterns and relationships between predictions across all prompts. We demonstrate the capabilities of KnowledgeVIS with feedback from six NLP experts as well as three different use cases: (1) probing biomedical knowledge in two domain-adapted models; and (2) evaluating harmful identity stereotypes and (3) discovering facts and relationships between three general-purpose models.
Collapse
|
7
|
Germain A, Sabol A, Chavali A, Fitzwilliams G, Cooper A, Khuon S, Green B, Kong C, Minna J, Kim YT. Machine learning enabled classification of lung cancer cell lines co-cultured with fibroblasts with lightweight convolutional neural network for initial diagnosis. J Biomed Sci 2024; 31:84. [PMID: 39180048 PMCID: PMC11344461 DOI: 10.1186/s12929-024-01071-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/28/2023] [Accepted: 08/11/2024] [Indexed: 08/26/2024] Open
Abstract
BACKGROUND Identification of lung cancer subtypes is critical for successful treatment in patients, especially those in advanced stages. Many advanced and personal treatments require knowledge of specific mutations, as well as up- and down-regulations of genes, for effective targeting of the cancer cells. While many studies focus on individual cell structures and delve deeper into gene sequencing, the present study proposes a machine learning method for lung cancer classification based on low-magnification cancer outgrowth patterns in a 2D co-culture environment. METHODS Using a magnetic well plate holder, circular pattern lung cancer cell clusters were generated among fibroblasts, and daily images were captured to monitor cancer outgrowth over a 9-day period. These outgrowth images were then augmented and used to train a convolutional neural network (CNN) model based on the lightweight TinyVGG architecture. The model was trained with pairs of classes representing three subtypes of NSCLC: A549 (adenocarcinoma), H520 (squamous cell carcinoma), and H460 (large cell carcinoma). The objective was to assess whether this lightweight machine learning model could accurately classify the three lung cancer cell lines at different stages of cancer outgrowth. Additionally, cancer outgrowth images of two patient-derived lung cancer cells, one with the KRAS oncogene and the other with the EGFR oncogene, were captured and classified using the CNN model. This demonstration aimed to investigate the translational potential of machine learning-enabled lung cancer classification. RESULTS The lightweight CNN model achieved over 93% classification accuracy at 1 day of outgrowth among A549, H460, and H520, and reached 100% classification accuracy at 7 days of outgrowth. Additionally, the model achieved 100% classification accuracy at 4 days for patient-derived lung cancer cells. Although these cells are classified as Adenocarcinoma, their outgrowth patterns vary depending on their oncogene expressions (KRAS or EGFR). CONCLUSIONS These results demonstrate that the lightweight CNN architecture, operating locally on a laptop without network or cloud connectivity, can effectively create a machine learning-enabled model capable of accurately classifying lung cancer cell subtypes, including those derived from patients, based upon their outgrowth patterns in the presence of surrounding fibroblasts. This advancement underscores the potential of machine learning to enhance early lung cancer subtyping, offering promising avenues for improving treatment outcomes in advanced stage-patients.
Collapse
Affiliation(s)
- Adam Germain
- Department of Bioengineering, University of Texas at Arlington, 500 UTA Blvd ERB244, Arlington, TX, 76010, USA
| | - Alex Sabol
- Department of Computer Science, University of Texas at Arlington, Arlington, TX, USA
| | - Anjani Chavali
- Department of Bioengineering, University of Texas at Arlington, 500 UTA Blvd ERB244, Arlington, TX, 76010, USA
| | - Giles Fitzwilliams
- Department of Bioengineering, University of Texas at Arlington, 500 UTA Blvd ERB244, Arlington, TX, 76010, USA
| | - Alexa Cooper
- Department of Biology, University of Texas at Arlington, Arlington, TX, USA
| | - Sandra Khuon
- Department of Nursing, University of Texas at Arlington, Arlington, TX, USA
| | - Bailey Green
- Department of Bioengineering, University of Texas at Arlington, 500 UTA Blvd ERB244, Arlington, TX, 76010, USA
| | - Calvin Kong
- Department of Bioengineering, University of Texas at Arlington, 500 UTA Blvd ERB244, Arlington, TX, 76010, USA
| | - John Minna
- Hamon Center for Therapeutic Oncology Research, University of Texas Southwestern Medical Center, Dallas, TX, USA
| | - Young-Tae Kim
- Department of Bioengineering, University of Texas at Arlington, 500 UTA Blvd ERB244, Arlington, TX, 76010, USA.
- Department of Urology, University of Texas Southwestern Medical Center, Dallas, TX, USA.
| |
Collapse
|
8
|
Alviz-Meza A, Hadechini-Meza LL, Peña-Ballesteros DY. Deep neural networks for external corrosion classification in industrial above-ground storage tanks. Heliyon 2024; 10:e34882. [PMID: 39959781 PMCID: PMC11829104 DOI: 10.1016/j.heliyon.2024.e34882] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 07/17/2024] [Accepted: 07/18/2024] [Indexed: 02/18/2025] Open
Abstract
This paper uses transfer learning and fine-tuning approaches to explore EfficientNet smaller models for a multiclassification task of corrosion types in above-ground storage tanks. Data augmentation was used to increment the data an oil and gas company provided, reaching a dataset of around 5000 images. The images were stored in Google Drive and imported into Colab to train the models using TensorFlow and Keras. After the hyperparameters' tuning a transfer learning model was selected and explored with fine-tuning. The EfficientNetB0 model delivered from fine-tuning accomplished 94 % performance. This work is the first attempt to deploy an artificial vision automatic tool for being implemented during tank inspection in the industrial sector. In further development, this model can be coupled with one based on object detection to remotely identify failures due to external corrosion issues, improving safety and reliability in oil and gas operations.
Collapse
Affiliation(s)
- Anibal Alviz-Meza
- Grupo de Investigación en Arquitectura, Diseño e Ingeniería (GIADI), Facultad de Ingeniería y Arquitectura, Institución Universitaria Mayor de Cartagena, Cartagena, Colombia
| | - Leidy L. Hadechini-Meza
- Grupo de Investigación Ciudad, Educación y Sociedad (CEUS), Facultad de Ciencias Sociales y Educación, Institución Universitaria Mayor de Cartagena, Cartagena, Colombia
| | - Darío Y. Peña-Ballesteros
- Grupo de Investigaciones en Corrosión (GIC), Universidad Industrial de Santander, Parque Tecnológico Guatiguará, Piedecuesta, 681011, Colombia
| |
Collapse
|
9
|
Sarmadi A, Razavi ZS, van Wijnen AJ, Soltani M. Comparative analysis of vision transformers and convolutional neural networks in osteoporosis detection from X-ray images. Sci Rep 2024; 14:18007. [PMID: 39097627 PMCID: PMC11297930 DOI: 10.1038/s41598-024-69119-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2024] [Accepted: 07/31/2024] [Indexed: 08/05/2024] Open
Abstract
Within the scope of this investigation, we carried out experiments to investigate the potential of the Vision Transformer (ViT) in the field of medical image analysis. The diagnosis of osteoporosis through inspection of X-ray radio-images is a substantial classification problem that we were able to address with the assistance of Vision Transformer models. In order to provide a basis for comparison, we conducted a parallel analysis in which we sought to solve the same problem by employing traditional convolutional neural networks (CNNs), which are well-known and commonly used techniques for the solution of image categorization issues. The findings of our research led us to conclude that ViT is capable of achieving superior outcomes compared to CNN. Furthermore, provided that methods have access to a sufficient quantity of training data, the probability increases that both methods arrive at more appropriate solutions to critical issues.
Collapse
Affiliation(s)
- Ali Sarmadi
- Department of Mechanical Engineering, K. N. Toosi University of Technology, Tehran, Iran
| | - Zahra Sadat Razavi
- Department of Mechanical Engineering, K. N. Toosi University of Technology, Tehran, Iran
- Physiology Research Center, Iran University Medical Sciences, Tehran, Iran
- Biochemistry Research Center, Iran University Medical Sciences, Tehran, Iran
| | - Andre J van Wijnen
- Department of Biochemistry, University of Vermont, Burlington, VT, USA
- Department of Internal Medicine, Erasmus University Medical Center, Rotterdam, The Netherlands
| | - Madjid Soltani
- Department of Mechanical Engineering, K. N. Toosi University of Technology, Tehran, Iran.
- Department of Electrical and Computer Engineering, University of Waterloo, Waterloo, Canada.
- Centre for Biotechnology and Bioengineering (CBB), University of Waterloo, Waterloo, Canada.
- Department of Integrative Oncology, BC Cancer Research Institute, Vancouver, Canada.
- Centre for Sustainable Business, International Business University, Toronto, Canada.
| |
Collapse
|
10
|
Fu Y, Stasko J. More Than Data Stories: Broadening the Role of Visualization in Contemporary Journalism. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2024; 30:5240-5259. [PMID: 37339040 DOI: 10.1109/tvcg.2023.3287585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/22/2023]
Abstract
Data visualization and journalism are deeply connected. From early infographics to recent data-driven storytelling, visualization has become an integrated part of contemporary journalism, primarily as a communication artifact to inform the general public. Data journalism, harnessing the power of data visualization, has emerged as a bridge between the growing volume of data and our society. Visualization research that centers around data storytelling has sought to understand and facilitate such journalistic endeavors. However, a recent metamorphosis in journalism has brought broader challenges and opportunities that extend beyond mere communication of data. We present this article to enhance our understanding of such transformations and thus broaden visualization research's scope and practical contribution to this evolving field. We first survey recent significant shifts, emerging challenges, and computational practices in journalism. We then summarize six roles of computing in journalism and their implications. Based on these implications, we provide propositions for visualization research concerning each role. Ultimately, by mapping the roles and propositions onto a proposed ecological model and contextualizing existing visualization research, we surface seven general topics and a series of research agendas that can guide future visualization research at this intersection.
Collapse
|
11
|
Cai YQ, Gong DX, Tang LY, Cai Y, Li HJ, Jing TC, Gong M, Hu W, Zhang ZW, Zhang X, Zhang GW. Pitfalls in Developing Machine Learning Models for Predicting Cardiovascular Diseases: Challenge and Solutions. J Med Internet Res 2024; 26:e47645. [PMID: 38869157 PMCID: PMC11316160 DOI: 10.2196/47645] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Revised: 10/30/2023] [Accepted: 06/12/2024] [Indexed: 06/14/2024] Open
Abstract
In recent years, there has been explosive development in artificial intelligence (AI), which has been widely applied in the health care field. As a typical AI technology, machine learning models have emerged with great potential in predicting cardiovascular diseases by leveraging large amounts of medical data for training and optimization, which are expected to play a crucial role in reducing the incidence and mortality rates of cardiovascular diseases. Although the field has become a research hot spot, there are still many pitfalls that researchers need to pay close attention to. These pitfalls may affect the predictive performance, credibility, reliability, and reproducibility of the studied models, ultimately reducing the value of the research and affecting the prospects for clinical application. Therefore, identifying and avoiding these pitfalls is a crucial task before implementing the research. However, there is currently a lack of a comprehensive summary on this topic. This viewpoint aims to analyze the existing problems in terms of data quality, data set characteristics, model design, and statistical methods, as well as clinical implications, and provide possible solutions to these problems, such as gathering objective data, improving training, repeating measurements, increasing sample size, preventing overfitting using statistical methods, using specific AI algorithms to address targeted issues, standardizing outcomes and evaluation criteria, and enhancing fairness and replicability, with the goal of offering reference and assistance to researchers, algorithm developers, policy makers, and clinical practitioners.
Collapse
Affiliation(s)
- Yu-Qing Cai
- The First Hospital of China Medical University, Shenyang, China
| | - Da-Xin Gong
- Smart Hospital Management Department, The First Hospital of China Medical University, Shenyang, China
| | - Li-Ying Tang
- The First Hospital of China Medical University, Shenyang, China
| | - Yue Cai
- The First Hospital of China Medical University, Shenyang, China
| | - Hui-Jun Li
- Shenyang Medical & Film Science and Technology Co, Ltd, Shenyang, China
| | - Tian-Ci Jing
- Smart Hospital Management Department, The First Hospital of China Medical University, Shenyang, China
| | | | - Wei Hu
- Bayi Orthopedic Hospital, Chengdu, China
| | - Zhen-Wei Zhang
- China Rongtong Medical & Healthcare Co, Ltd, Chengdu, China
| | - Xingang Zhang
- Department of Cardiology, The First Hospital of China Medical University, Shenyang, China
| | - Guang-Wei Zhang
- Smart Hospital Management Department, The First Hospital of China Medical University, Shenyang, China
| |
Collapse
|
12
|
Cai Y, Cai YQ, Tang LY, Wang YH, Gong M, Jing TC, Li HJ, Li-Ling J, Hu W, Yin Z, Gong DX, Zhang GW. Artificial intelligence in the risk prediction models of cardiovascular disease and development of an independent validation screening tool: a systematic review. BMC Med 2024; 22:56. [PMID: 38317226 PMCID: PMC10845808 DOI: 10.1186/s12916-024-03273-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Accepted: 01/23/2024] [Indexed: 02/07/2024] Open
Abstract
BACKGROUND A comprehensive overview of artificial intelligence (AI) for cardiovascular disease (CVD) prediction and a screening tool of AI models (AI-Ms) for independent external validation are lacking. This systematic review aims to identify, describe, and appraise AI-Ms of CVD prediction in the general and special populations and develop a new independent validation score (IVS) for AI-Ms replicability evaluation. METHODS PubMed, Web of Science, Embase, and IEEE library were searched up to July 2021. Data extraction and analysis were performed for the populations, distribution, predictors, algorithms, etc. The risk of bias was evaluated with the prediction risk of bias assessment tool (PROBAST). Subsequently, we designed IVS for model replicability evaluation with five steps in five items, including transparency of algorithms, performance of models, feasibility of reproduction, risk of reproduction, and clinical implication, respectively. The review is registered in PROSPERO (No. CRD42021271789). RESULTS In 20,887 screened references, 79 articles (82.5% in 2017-2021) were included, which contained 114 datasets (67 in Europe and North America, but 0 in Africa). We identified 486 AI-Ms, of which the majority were in development (n = 380), but none of them had undergone independent external validation. A total of 66 idiographic algorithms were found; however, 36.4% were used only once and only 39.4% over three times. A large number of different predictors (range 5-52,000, median 21) and large-span sample size (range 80-3,660,000, median 4466) were observed. All models were at high risk of bias according to PROBAST, primarily due to the incorrect use of statistical methods. IVS analysis confirmed only 10 models as "recommended"; however, 281 and 187 were "not recommended" and "warning," respectively. CONCLUSION AI has led the digital revolution in the field of CVD prediction, but is still in the early stage of development as the defects of research design, report, and evaluation systems. The IVS we developed may contribute to independent external validation and the development of this field.
Collapse
Affiliation(s)
- Yue Cai
- China Medical University, Shenyang, 110122, China
| | - Yu-Qing Cai
- China Medical University, Shenyang, 110122, China
| | - Li-Ying Tang
- China Medical University, Shenyang, 110122, China
| | - Yi-Han Wang
- China Medical University, Shenyang, 110122, China
| | - Mengchun Gong
- Digital Health China Co. Ltd, Beijing, 100089, China
| | - Tian-Ci Jing
- Smart Hospital Management Department, the First Hospital of China Medical University, Shenyang, 110001, China
| | - Hui-Jun Li
- Shenyang Medical & Film Science and Technology Co. Ltd., Shenyang, 110001, China
- Enduring Medicine Smart Innovation Research Institute, Shenyang, 110001, China
| | - Jesse Li-Ling
- Institute of Genetic Medicine, School of Life Science, State Key Laboratory of Biotherapy, Sichuan University, Chengdu, 610065, China
| | - Wei Hu
- Bayi Orthopedic Hospital, Chengdu, 610017, China
| | - Zhihua Yin
- Department of Epidemiology, School of Public Health, China Medical University, Shenyang, 110122, China.
| | - Da-Xin Gong
- Smart Hospital Management Department, the First Hospital of China Medical University, Shenyang, 110001, China.
- The Internet Hospital Branch of the Chinese Research Hospital Association, Beijing, 100006, China.
| | - Guang-Wei Zhang
- Smart Hospital Management Department, the First Hospital of China Medical University, Shenyang, 110001, China.
- The Internet Hospital Branch of the Chinese Research Hospital Association, Beijing, 100006, China.
| |
Collapse
|
13
|
Dou H, Shen F, Zhao J, Mu X. Understanding neural network through neuron level visualization. Neural Netw 2023; 168:484-495. [PMID: 37806141 DOI: 10.1016/j.neunet.2023.09.030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2023] [Revised: 08/09/2023] [Accepted: 09/17/2023] [Indexed: 10/10/2023]
Abstract
Neurons are the fundamental units of neural networks. In this paper, we propose a method for explaining neural networks by visualizing the learning process of neurons. For a trained neural network, the proposed method obtains the features learned by each neuron and displays the features in a human-understandable form. The features learned by different neurons are combined to analyze the working mechanism of different neural network models. The method is applicable to neural networks without requiring any changes to the architectures of the models. In this study, we apply the proposed method to both Fully Connected Networks (FCNs) and Convolutional Neural Networks (CNNs) trained using the backpropagation learning algorithm. We conduct experiments on models for image classification tasks to demonstrate the effectiveness of the method. Through these experiments, we gain insights into the working mechanisms of various neural network architectures and evaluate neural network interpretability from diverse perspectives.
Collapse
Affiliation(s)
- Hui Dou
- State Key Laboratory for Novel Software Technology, China; Department of Computer Science and Technology, Nanjing University, Nanjing 210023, China.
| | - Furao Shen
- State Key Laboratory for Novel Software Technology, China; School of Artificial Intelligence, Nanjing University, Nanjing 210023, China.
| | - Jian Zhao
- School of Electronic Science and Engineering, Nanjing University, Nanjing 210023, China.
| | - Xinyu Mu
- State Key Laboratory for Novel Software Technology, China; School of Artificial Intelligence, Nanjing University, Nanjing 210023, China
| |
Collapse
|
14
|
Wang S, Ren Y, Xia B, Liu K, Li H. Prediction of atmospheric pollutants in urban environment based on coupled deep learning model and sensitivity analysis. CHEMOSPHERE 2023; 331:138830. [PMID: 37137395 DOI: 10.1016/j.chemosphere.2023.138830] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Revised: 04/11/2023] [Accepted: 04/30/2023] [Indexed: 05/05/2023]
Abstract
Accurate and efficient predictions of pollutants in the atmosphere provide a reliable basis for the scientific management of atmospheric pollution. This study develops a model that combines an attention mechanism, convolutional neural network (CNN), and long short-term memory (LSTM) unit to predict the O3 and PM2.5 levels in the atmosphere, as well as an air quality index (AQI). The prediction results given by the proposed model are compared with those from CNN-LSTM and LSTM models as well as random forest and support vector regression models. The proposed model achieves a correlation coefficient between the predicted and observed values of more than 0.90, outperforming the other four models. The model errors are also consistently lower when using the proposed approach. Sobol-based sensitivity analysis is applied to identify the variables that make the greatest contribution to the model prediction results. Taking the COVID-19 outbreak as the time boundary, we find some homology in the interactions among the pollutants and meteorological factors in the atmosphere during different periods. Solar irradiance is the most important factor for O3, CO is the most important factor for PM2.5, and particulate matter has the most significant effect on AQI. The key influencing factors are the same over the whole phase and before the COVID-19 outbreak, indicating that the impact of COVID-19 restrictions on AQI gradually stabilized. Removing variables that contribute the least to the prediction results without affecting the model prediction performance improves the modeling efficiency and reduces the computational costs.
Collapse
Affiliation(s)
- Siyuan Wang
- School of Environment, Nanjing Normal University, Nanjing, 210023, PR China; School of Mathematics and Computer Science, Yan'an University, Yan'an, 716000, PR China
| | - Ying Ren
- School of Mathematics and Computer Science, Yan'an University, Yan'an, 716000, PR China
| | - Bisheng Xia
- School of Mathematics and Computer Science, Yan'an University, Yan'an, 716000, PR China
| | - Kai Liu
- School of Environment, Nanjing Normal University, Nanjing, 210023, PR China
| | - Huiming Li
- School of Environment, Nanjing Normal University, Nanjing, 210023, PR China.
| |
Collapse
|
15
|
Yang S, Yang Z, Yang J. 4mCBERT: A computing tool for the identification of DNA N4-methylcytosine sites by sequence- and chemical-derived information based on ensemble learning strategies. Int J Biol Macromol 2023; 231:123180. [PMID: 36646347 DOI: 10.1016/j.ijbiomac.2023.123180] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2022] [Revised: 11/26/2022] [Accepted: 12/30/2022] [Indexed: 01/15/2023]
Abstract
N4-methylcytosine (4mC) is an important DNA chemical modification pattern which is a new methylation modification discovered in recent years and plays critical roles in gene expression regulation, defense against invading genetic elements, genomic imprinting, and so on. Identifying 4mC site from DNA sequence segment contributes to discovering more novel modification patterns. In this paper, we present a model called 4mCBERT that encodes DNA sequence segments by sequence characteristics including one-hot, electron-ion interaction pseudopotential, nucleotide chemical property, word2vec and chemical information containing physicochemical properties (PCP), chemical bidirectional encoder representations from transformers (chemical BERT) and employs ensemble learning framework to develop a prediction model. PCP and chemical BERT features are firstly constructed and applied to predict 4mC sites and show positive contributions to identifying 4mC. For the Matthew's Correlation Coefficient, 4mCBERT significantly outperformed other state-of-the-art models on six independent benchmark datasets including A. thaliana, C. elegans, D. melanogaster, E. coli, G. Pickering, and G. subterraneous by 4.32 % to 24.39 %, 2.52 % to 31.65 %, 2 % to 16.49 %, 6.63 % to 35.15, 8.59 % to 61.85 %, and 8.45 % to 34.45 %. Moreover, 4mCBERT is designed to allow users to predict 4mC sites and retrain 4mC prediction models. In brief, 4mCBERT shows higher performance on six benchmark datasets by incorporating sequence- and chemical-driven information and is available at http://cczubio.top/4mCBERT and https://github.com/abcair/4mCBERT.
Collapse
Affiliation(s)
- Sen Yang
- School of Computer Science and Artificial Intelligence, Aliyun School of Big Data, School of Software, Changzhou 213164, China; The Affiliated Changzhou No 2 People's Hospital of Nanjing Medical University, Changzhou 213164, China.
| | - Zexi Yang
- School of Computer Science and Artificial Intelligence, Aliyun School of Big Data, School of Software, Changzhou 213164, China
| | - Jun Yang
- School of Educational Sciences, Yili Normal University, Yining 835000, China
| |
Collapse
|
16
|
Yuan J, Liu M, Tian F, Liu S. Visual Analysis of Neural Architecture Spaces for Summarizing Design Principles. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2023; 29:288-298. [PMID: 36191103 DOI: 10.1109/tvcg.2022.3209404] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Recent advances in artificial intelligence largely benefit from better neural network architectures. These architectures are a product of a costly process of trial-and-error. To ease this process, we develop ArchExplorer, a visual analysis method for understanding a neural architecture space and summarizing design principles. The key idea behind our method is to make the architecture space explainable by exploiting structural distances between architectures. We formulate the pairwise distance calculation as solving an all-pairs shortest path problem. To improve efficiency, we decompose this problem into a set of single-source shortest path problems. The time complexity is reduced from O(kn2N) to O(knN). Architectures are hierarchically clustered according to the distances between them. A circle-packing-based architecture visualization has been developed to convey both the global relationships between clusters and local neighborhoods of the architectures in each cluster. Two case studies and a post-analysis are presented to demonstrate the effectiveness of ArchExplorer in summarizing design principles and selecting better-performing architectures.
Collapse
|
17
|
Goh HA, Ho CK, Abas FS. Front-end deep learning web apps development and deployment: a review. APPL INTELL 2022; 53:15923-15945. [PMID: 36466774 PMCID: PMC9709375 DOI: 10.1007/s10489-022-04278-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/17/2022] [Indexed: 12/03/2022]
Abstract
Machine learning and deep learning models are commonly developed using programming languages such as Python, C++, or R and deployed as web apps delivered from a back-end server or as mobile apps installed from an app store. However, recently front-end technologies and JavaScript libraries, such as TensorFlow.js, have been introduced to make machine learning more accessible to researchers and end-users. Using JavaScript, TensorFlow.js can define, train, and run new or existing, pre-trained machine learning models entirely in the browser from the client-side, which improves the user experience through interaction while preserving privacy. Deep learning models deployed on front-end browsers must be small, have fast inference, and ideally be interactive in real-time. Therefore, the emphasis on development and deployment is different. This paper aims to review the development and deployment of these deep-learning web apps to raise awareness of the recent advancements and encourage more researchers to take advantage of this technology for their own work. First, the rationale behind the deployment stack (front-end, JavaScript, and TensorFlow.js) is discussed. Then, the development approach for obtaining deep learning models that are optimized and suitable for front-end deployment is then described. The article also provides current web applications divided into seven categories to show deep learning potential on the front end. These include web apps for deep learning playground, pose detection and gesture tracking, music and art creation, expression detection and facial recognition, video segmentation, image and signal analysis, healthcare diagnosis, recognition, and identification.
Collapse
Affiliation(s)
- Hock-Ann Goh
- Faculty of Engineering and Technology, Multimedia University, Jalan Ayer Keroh Lama, Bukit Beruang, 75450 Melaka Malaysia
| | - Chin-Kuan Ho
- Asia Pacific University of Technology and Innovation, Jalan Teknologi 5, Technology Park Malaysia, 57000 Kuala Lumpur, Malaysia
| | - Fazly Salleh Abas
- Faculty of Engineering and Technology, Multimedia University, Jalan Ayer Keroh Lama, Bukit Beruang, 75450 Melaka Malaysia
| |
Collapse
|
18
|
Touretzky D, Gardner-McCune C, Seehorn D. Machine Learning and the Five Big Ideas in AI. INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION 2022. [DOI: 10.1007/s40593-022-00314-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
19
|
Xiang J, Pan R, Gao W. An Efficient Retrieval System Framework for Fabrics Based on Fine-Grained Similarity. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1319. [PMID: 36141205 PMCID: PMC9497872 DOI: 10.3390/e24091319] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/27/2022] [Revised: 08/16/2022] [Accepted: 08/17/2022] [Indexed: 06/16/2023]
Abstract
In the context of "double carbon", as a traditional high energy consumption industry, the textile industry is facing the severe challenges of energy saving and emission reduction. To improve production efficiency in the textile industry, we propose the use of content-based image retrieval technology to shorten the fabric production cycle. However, fabric retrieval has high requirements for results, which makes it difficult for common retrieval methods to be directly applied to fabric retrieval. This paper presents a novel method for fabric image retrieval. Firstly, we define a fine-grained similarity to measure the similarity between two fabric images. Then, a convolutional neural network with a compact structure and cross-domain connections is designed to narrow the gap between fabric images and similarities. To overcome the problems of probabilistic missing and difficult training in classical hashing, we introduce a variational network module and structural module into the hashing model, which is called DVSH. We employ list-wise learning to perform similarity embedding. The experimental results demonstrate the superiority and efficiency of the proposed hashing model, DVSH.
Collapse
|
20
|
Sign and Human Action Detection Using Deep Learning. J Imaging 2022; 8:jimaging8070192. [PMID: 35877636 PMCID: PMC9316580 DOI: 10.3390/jimaging8070192] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2022] [Revised: 04/05/2022] [Accepted: 04/16/2022] [Indexed: 02/04/2023] Open
Abstract
Human beings usually rely on communication to express their feeling and ideas and to solve disputes among themselves. A major component required for effective communication is language. Language can occur in different forms, including written symbols, gestures, and vocalizations. It is usually essential for all of the communicating parties to be fully conversant with a common language. However, to date this has not been the case between speech-impaired people who use sign language and people who use spoken languages. A number of different studies have pointed out a significant gaps between these two groups which can limit the ease of communication. Therefore, this study aims to develop an efficient deep learning model that can be used to predict British sign language in an attempt to narrow this communication gap between speech-impaired and non-speech-impaired people in the community. Two models were developed in this research, CNN and LSTM, and their performance was evaluated using a multi-class confusion matrix. The CNN model emerged with the highest performance, attaining training and testing accuracies of 98.8% and 97.4%, respectively. In addition, the model achieved average weighted precession and recall of 97% and 96%, respectively. On the other hand, the LSTM model’s performance was quite poor, with the maximum training and testing performance accuracies achieved being 49.4% and 48.7%, respectively. Our research concluded that the CNN model was the best for recognizing and determining British sign language.
Collapse
|
21
|
Fei Q. Analysis of Traditional Cultural Acceptance Based on Deep Learning. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:2373842. [PMID: 35707191 PMCID: PMC9192241 DOI: 10.1155/2022/2373842] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/10/2022] [Revised: 04/08/2022] [Accepted: 04/13/2022] [Indexed: 11/18/2022]
Abstract
Technological development has resulted in the utilisation of advanced technology search using deep learning technology in all industries. Artificial intelligence is set to be filled with machines that can perform tasks that need human intelligence. These machines are controlled by computer systems. In this study, people are going to analyse the traditional cultural acceptance based on deep learning. Deep learning technology is a subfield of machine learning that makes use of algorithms that are inspired by the human brain. Artificial neural network is one such network used in deep learning. Traditional cultural acceptance is an integration of traditional cultures and cultural acceptance. The traditional cultures are said to be the tribes that are never exposed to the technological radar. These communities live traditionally in remote areas where there is no access to the latest technologies. The traditional cultures were found in the upper Amazon regions. Because they have not been exposed to the modern world, they have retained their ancient traditional and cultural value. Cultural acceptance can be said in simple words: accepting people as they are, without shaming their culture, traditional habits, attire, and body. In this proposed system, deep learning techniques were used to analyse the traditional cultural acceptance. The results proved that the proposed model called Faster R-CNN performs well than the existing algorithms.
Collapse
Affiliation(s)
- Qingmei Fei
- Sichuan Vocational College of Cultural Industries, Chengdu 610000l, Sichuan, China
| |
Collapse
|
22
|
Hurtado Sánchez JA, Casilimas K, Caicedo Rendon OM. Deep Reinforcement Learning for Resource Management on Network Slicing: A Survey. SENSORS 2022; 22:s22083031. [PMID: 35459015 PMCID: PMC9032530 DOI: 10.3390/s22083031] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/01/2022] [Revised: 04/08/2022] [Accepted: 04/11/2022] [Indexed: 12/04/2022]
Abstract
Network Slicing and Deep Reinforcement Learning (DRL) are vital enablers for achieving 5G and 6G networks. A 5G/6G network can comprise various network slices from unique or multiple tenants. Network providers need to perform intelligent and efficient resource management to offer slices that meet the quality of service and quality of experience requirements of 5G/6G use cases. Resource management is far from being a straightforward task. This task demands complex and dynamic mechanisms to control admission and allocate, schedule, and orchestrate resources. Intelligent and effective resource management needs to predict the services’ demand coming from tenants (each tenant with multiple network slice requests) and achieve autonomous behavior of slices. This paper identifies the relevant phases for resource management in network slicing and analyzes approaches using reinforcement learning (RL) and DRL algorithms for realizing each phase autonomously. We analyze the approaches according to the optimization objective, the network focus (core, radio access, edge, and end-to-end network), the space of states, the space of actions, the algorithms, the structure of deep neural networks, the exploration–exploitation method, and the use cases (or vertical applications). We also provide research directions related to RL/DRL-based network slice resource management.
Collapse
|
23
|
Nijiati M, Ma J, Hu C, Tuersun A, Abulizi A, Kelimu A, Zhang D, Li G, Zou X. Artificial Intelligence Assisting the Early Detection of Active Pulmonary Tuberculosis From Chest X-Rays: A Population-Based Study. Front Mol Biosci 2022; 9:874475. [PMID: 35463963 PMCID: PMC9023793 DOI: 10.3389/fmolb.2022.874475] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2022] [Accepted: 03/08/2022] [Indexed: 11/13/2022] Open
Abstract
As a major infectious disease, tuberculosis (TB) still poses a threat to people’s health in China. As a triage test for TB, reading chest radiography with traditional approach ends up with high inter-radiologist and intra-radiologist variability, moderate specificity and a waste of time and medical resources. Thus, this study established a deep convolutional neural network (DCNN) based artificial intelligence (AI) algorithm, aiming at diagnosing TB on posteroanterior chest X-ray photographs in an effective and accurate way. Altogether, 5,000 patients with TB and 4,628 patients without TB were included in the study, totaling to 9,628 chest X-ray photographs analyzed. Splitting the radiographs into a training set (80.4%) and a testing set (19.6%), three different DCNN algorithms, including ResNet, VGG, and AlexNet, were trained to classify the chest radiographs as images of pulmonary TB or without TB. Both the diagnostic accuracy and the area under the receiver operating characteristic curve were used to evaluate the performance of the three AI diagnosis models. Reaching an accuracy of 96.73% and marking the precise TB regions on the radiographs, ResNet algorithm-based AI outperformed the rest models and showed excellent diagnostic ability in different clinical subgroups in the stratification analysis. In summary, the ResNet algorithm-based AI diagnosis system provided accurate TB diagnosis, which could have broad prospects in clinical application for TB diagnosis, especially in poor regions with high TB incidence.
Collapse
Affiliation(s)
- Mayidili Nijiati
- Department of Radiology, The First People’s Hospital of Kashi Prefecture, Kashi, China
- *Correspondence: Mayidili Nijiati, ; Guanbin Li, ; Xiaoguang Zou,
| | - Jie Ma
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, China
| | - Chuling Hu
- Department of Colorectal Surgery, The Sixth Affiliated Hospital, Sun Yat-sen University, Guangzhou, China
| | - Abudouresuli Tuersun
- Department of Radiology, The First People’s Hospital of Kashi Prefecture, Kashi, China
| | | | - Abudoureyimu Kelimu
- Department of Radiology, Kashi Area Tuberculosis Control Center, Kashi, China
| | - Dongyu Zhang
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, China
| | - Guanbin Li
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, China
- *Correspondence: Mayidili Nijiati, ; Guanbin Li, ; Xiaoguang Zou,
| | - Xiaoguang Zou
- Clinical Medical Research Center, The First People’s Hospital of Kashi Prefecture, Kashi, China
- *Correspondence: Mayidili Nijiati, ; Guanbin Li, ; Xiaoguang Zou,
| |
Collapse
|
24
|
Yuan H, Lee JH, Zhang S. Research on simulation of 3D human animation vision technology based on an enhanced machine learning algorithm. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07083-x] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
25
|
Smith JT, Ochoa M, Faulkner D, Haskins G, Intes X. Deep learning in macroscopic diffuse optical imaging. JOURNAL OF BIOMEDICAL OPTICS 2022; 27:JBO-210288VRR. [PMID: 35218169 PMCID: PMC8881080 DOI: 10.1117/1.jbo.27.2.020901] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Accepted: 02/09/2022] [Indexed: 05/02/2023]
Abstract
SIGNIFICANCE Biomedical optics system design, image formation, and image analysis have primarily been guided by classical physical modeling and signal processing methodologies. Recently, however, deep learning (DL) has become a major paradigm in computational modeling and has demonstrated utility in numerous scientific domains and various forms of data analysis. AIM We aim to comprehensively review the use of DL applied to macroscopic diffuse optical imaging (DOI). APPROACH First, we provide a layman introduction to DL. Then, the review summarizes current DL work in some of the most active areas of this field, including optical properties retrieval, fluorescence lifetime imaging, and diffuse optical tomography. RESULTS The advantages of using DL for DOI versus conventional inverse solvers cited in the literature reviewed herein are numerous. These include, among others, a decrease in analysis time (often by many orders of magnitude), increased quantitative reconstruction quality, robustness to noise, and the unique capability to learn complex end-to-end relationships. CONCLUSIONS The heavily validated capability of DL's use across a wide range of complex inverse solving methodologies has enormous potential to bring novel DOI modalities, otherwise deemed impractical for clinical translation, to the patient's bedside.
Collapse
Affiliation(s)
- Jason T. Smith
- Rensselaer Polytechnic Institute, Department of Biomedical Engineering, Troy, New York, United States
| | - Marien Ochoa
- Rensselaer Polytechnic Institute, Department of Biomedical Engineering, Troy, New York, United States
| | - Denzel Faulkner
- Rensselaer Polytechnic Institute, Department of Biomedical Engineering, Troy, New York, United States
| | - Grant Haskins
- Rensselaer Polytechnic Institute, Department of Biomedical Engineering, Troy, New York, United States
| | - Xavier Intes
- Rensselaer Polytechnic Institute, Center for Modeling, Simulation and Imaging for Medicine, Troy, New York, United States
| |
Collapse
|
26
|
Energy Management Strategy Based on a Novel Speed Prediction Method. SENSORS 2021; 21:s21248273. [PMID: 34960362 PMCID: PMC8707142 DOI: 10.3390/s21248273] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/17/2021] [Revised: 12/04/2021] [Accepted: 12/09/2021] [Indexed: 11/21/2022]
Abstract
Vehicle speed prediction can obtain the future driving status of a vehicle in advance, which helps to make better decisions for energy management strategies. We propose a novel deep learning neural network architecture for vehicle speed prediction, called VSNet, by combining convolutional neural network (CNN) and long-short term memory network (LSTM). VSNet adopts a fake image composed of 15 vehicle signals in the past 15 s as model input to predict the vehicle speed in the next 5 s. Different from the traditional series or parallel structure, VSNet is structured with CNN and LSTM in series and then in parallel with two other CNNs of different convolutional kernel sizes. The unique architecture allows for better fitting of highly nonlinear relationships. The prediction performance of VSNet is first examined. The prediction results show a RMSE range of 0.519–2.681 and a R2 range of 0.997–0.929 for the future 5 s. Finally, an energy management strategy combined with VSNet and model predictive control (MPC) is simulated. The equivalent fuel consumption of the simulation increases by only 4.74% compared with DP-based energy management strategy and decreased by 2.82% compared with the speed prediction method with low accuracy.
Collapse
|
27
|
Abstract
AbstractWe investigate the capabilities of transfer learning in the area of structural health monitoring. In particular, we are interested in damage detection for concrete structures. Typical image datasets for such problems are relatively small, calling for the transfer of learned representation from a related large-scale dataset. Past efforts of damage detection using images have mainly considered cross-domain transfer learning approaches using pre-trained ImageNet models that are subsequently fine-tuned for the target task. However, there are rising concerns about the generalizability of ImageNet representations for specific target domains, such as for visual inspection and medical imaging. We, therefore, evaluate a combination of in-domain and cross-domain transfer learning strategies for damage detection in bridges. We perform comprehensive comparisons to study the impact of cross-domain and in-domain transfer, with various initialization strategies, using six publicly available visual inspection datasets. The pre-trained models are also evaluated for their ability to cope with the extremely low-data regime. We show that the combination of cross-domain and in-domain transfer persistently shows superior performance specially with tiny datasets. Likewise, we also provide visual explanations of predictive models to enable algorithmic transparency and provide insights to experts about the intrinsic decision logic of typically black-box deep models.
Collapse
|
28
|
Jiang T, Hu XJ, Yao XH, Tu LP, Huang JB, Ma XX, Cui J, Wu QF, Xu JT. Tongue image quality assessment based on a deep convolutional neural network. BMC Med Inform Decis Mak 2021; 21:147. [PMID: 33952228 PMCID: PMC8097848 DOI: 10.1186/s12911-021-01508-8] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2020] [Accepted: 04/27/2021] [Indexed: 12/26/2022] Open
Abstract
BACKGROUND Tongue diagnosis is an important research field of TCM diagnostic technology modernization. The quality of tongue images is the basis for constructing a standard dataset in the field of tongue diagnosis. To establish a standard tongue image database in the TCM industry, we need to evaluate the quality of a massive number of tongue images and add qualified images to the database. Therefore, an automatic, efficient and accurate quality control model is of significance to the development of intelligent tongue diagnosis technology for TCM. METHODS Machine learning methods, including Support Vector Machine (SVM), Random Forest (RF), Gradient Boosting Decision Tree (GBDT), Adaptive Boosting Algorithm (Adaboost), Naïve Bayes, Decision Tree (DT), Residual Neural Network (ResNet), Convolution Neural Network developed by Visual Geometry Group at University of Oxford (VGG), and Densely Connected Convolutional Networks (DenseNet), were utilized to identify good-quality and poor-quality tongue images. Their performances were made comparisons by using metrics such as accuracy, precision, recall, and F1-Score. RESULTS The experimental results showed that the accuracy of the three deep learning models was more than 96%, and the accuracy of ResNet-152 and DenseNet-169 was more than 98%. The model ResNet-152 obtained accuracy of 99.04%, precision of 99.05%, recall of 99.04%, and F1-score of 99.05%. The performances were better than performances of other eight models. The eight models are VGG-16, DenseNet-169, SVM, RF, GBDT, Adaboost, Naïve Bayes, and DT. ResNet-152 was selected as quality-screening model for tongue IQA. CONCLUSIONS Our research findings demonstrate various CNN models in the decision-making process for the selection of tongue image quality assessment and indicate that applying deep learning methods, specifically deep CNNs, to evaluate poor-quality tongue images is feasible.
Collapse
Affiliation(s)
- Tao Jiang
- Basic Medical College Shanghai University of Traditional Chinese Medicine, 1200 Cailun Road, Pudong New Area, Shanghai, 201203, China
| | - Xiao-Juan Hu
- Shanghai Collaborative Innovation Center of Health Service in TCM, Shanghai University of TCM, 1200 Cailun Road, Shanghai, 201203, China
| | - Xing-Hua Yao
- Basic Medical College Shanghai University of Traditional Chinese Medicine, 1200 Cailun Road, Pudong New Area, Shanghai, 201203, China
| | - Li-Ping Tu
- Basic Medical College Shanghai University of Traditional Chinese Medicine, 1200 Cailun Road, Pudong New Area, Shanghai, 201203, China
| | - Jing-Bin Huang
- Basic Medical College Shanghai University of Traditional Chinese Medicine, 1200 Cailun Road, Pudong New Area, Shanghai, 201203, China
| | - Xu-Xiang Ma
- Basic Medical College Shanghai University of Traditional Chinese Medicine, 1200 Cailun Road, Pudong New Area, Shanghai, 201203, China
| | - Ji Cui
- Basic Medical College Shanghai University of Traditional Chinese Medicine, 1200 Cailun Road, Pudong New Area, Shanghai, 201203, China
| | - Qing-Feng Wu
- School of Information Science and Engineering, Xiamen University, Xiamen, 361005, China
| | - Jia-Tuo Xu
- Basic Medical College Shanghai University of Traditional Chinese Medicine, 1200 Cailun Road, Pudong New Area, Shanghai, 201203, China.
| |
Collapse
|
29
|
Artificial intelligence in cyber security: research advances, challenges, and opportunities. Artif Intell Rev 2021. [DOI: 10.1007/s10462-021-09976-0] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/21/2023]
|
30
|
Neuroscope: An Explainable AI Toolbox for Semantic Segmentation and Image Classification of Convolutional Neural Nets. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11052199] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Trust in artificial intelligence (AI) predictions is a crucial point for a widespread acceptance of new technologies, especially in sensitive areas like autonomous driving. The need for tools explaining AI for deep learning of images is thus eminent. Our proposed toolbox Neuroscope addresses this demand by offering state-of-the-art visualization algorithms for image classification and newly adapted methods for semantic segmentation of convolutional neural nets (CNNs). With its easy to use graphical user interface (GUI), it provides visualization on all layers of a CNN. Due to its open model-view-controller architecture, networks generated and trained with Keras and PyTorch are processable, with an interface allowing extension to additional frameworks. We demonstrate the explanation abilities provided by Neuroscope using the example of traffic scene analysis.
Collapse
|
31
|
Muralitharan S, Nelson W, Di S, McGillion M, Devereaux PJ, Barr NG, Petch J. Machine Learning-Based Early Warning Systems for Clinical Deterioration: Systematic Scoping Review. J Med Internet Res 2021; 23:e25187. [PMID: 33538696 PMCID: PMC7892287 DOI: 10.2196/25187] [Citation(s) in RCA: 53] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2020] [Revised: 12/19/2020] [Accepted: 12/20/2020] [Indexed: 01/04/2023] Open
Abstract
BACKGROUND Timely identification of patients at a high risk of clinical deterioration is key to prioritizing care, allocating resources effectively, and preventing adverse outcomes. Vital signs-based, aggregate-weighted early warning systems are commonly used to predict the risk of outcomes related to cardiorespiratory instability and sepsis, which are strong predictors of poor outcomes and mortality. Machine learning models, which can incorporate trends and capture relationships among parameters that aggregate-weighted models cannot, have recently been showing promising results. OBJECTIVE This study aimed to identify, summarize, and evaluate the available research, current state of utility, and challenges with machine learning-based early warning systems using vital signs to predict the risk of physiological deterioration in acutely ill patients, across acute and ambulatory care settings. METHODS PubMed, CINAHL, Cochrane Library, Web of Science, Embase, and Google Scholar were searched for peer-reviewed, original studies with keywords related to "vital signs," "clinical deterioration," and "machine learning." Included studies used patient vital signs along with demographics and described a machine learning model for predicting an outcome in acute and ambulatory care settings. Data were extracted following PRISMA, TRIPOD, and Cochrane Collaboration guidelines. RESULTS We identified 24 peer-reviewed studies from 417 articles for inclusion; 23 studies were retrospective, while 1 was prospective in nature. Care settings included general wards, intensive care units, emergency departments, step-down units, medical assessment units, postanesthetic wards, and home care. Machine learning models including logistic regression, tree-based methods, kernel-based methods, and neural networks were most commonly used to predict the risk of deterioration. The area under the curve for models ranged from 0.57 to 0.97. CONCLUSIONS In studies that compared performance, reported results suggest that machine learning-based early warning systems can achieve greater accuracy than aggregate-weighted early warning systems but several areas for further research were identified. While these models have the potential to provide clinical decision support, there is a need for standardized outcome measures to allow for rigorous evaluation of performance across models. Further research needs to address the interpretability of model outputs by clinicians, clinical efficacy of these systems through prospective study design, and their potential impact in different clinical settings.
Collapse
Affiliation(s)
- Sankavi Muralitharan
- Centre for Data Science and Digital Health, Hamilton Health Sciences, Hamilton, ON, Canada
- DeGroote School of Business, McMaster University, Hamilton, ON, Canada
| | - Walter Nelson
- Centre for Data Science and Digital Health, Hamilton Health Sciences, Hamilton, ON, Canada
| | - Shuang Di
- Centre for Data Science and Digital Health, Hamilton Health Sciences, Hamilton, ON, Canada
- Dalla Lana School of Public Health, University of Toronto, Toronto, ON, Canada
| | - Michael McGillion
- School of Nursing, McMaster University, Hamilton, ON, Canada
- Population Health Research Institute, Hamilton, ON, Canada
| | - P J Devereaux
- Population Health Research Institute, Hamilton, ON, Canada
- Departments of Health Evidence and Impact and Medicine, McMaster University, Hamilton, ON, Canada
| | - Neil Grant Barr
- Health Policy and Management, DeGroote School of Business, McMaster University, Hamilton, ON, Canada
| | - Jeremy Petch
- Centre for Data Science and Digital Health, Hamilton Health Sciences, Hamilton, ON, Canada
- Population Health Research Institute, Hamilton, ON, Canada
- Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, ON, Canada
- Department of Medicine, Faculty of Health Sciences, McMaster University, Hamilton, ON, Canada
| |
Collapse
|