1
|
Swaminathan A, Lopez I, Wang W, Srivastava U, Tran E, Bhargava-Shah A, Wu JY, Ren AL, Caoili K, Bui B, Alkhani L, Lee S, Mohit N, Seo N, Macedo N, Cheng W, Liu C, Thomas R, Chen JH, Gevaert O. Selective prediction for extracting unstructured clinical data. J Am Med Inform Assoc 2023; 31:188-197. [PMID: 37769323 PMCID: PMC10746316 DOI: 10.1093/jamia/ocad182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Revised: 08/21/2023] [Accepted: 08/24/2023] [Indexed: 09/30/2023] Open
Abstract
OBJECTIVE While there are currently approaches to handle unstructured clinical data, such as manual abstraction and structured proxy variables, these methods may be time-consuming, not scalable, and imprecise. This article aims to determine whether selective prediction, which gives a model the option to abstain from generating a prediction, can improve the accuracy and efficiency of unstructured clinical data abstraction. MATERIALS AND METHODS We trained selective classifiers (logistic regression, random forest, support vector machine) to extract 5 variables from clinical notes: depression (n = 1563), glioblastoma (GBM, n = 659), rectal adenocarcinoma (DRA, n = 601), and abdominoperineal resection (APR, n = 601) and low anterior resection (LAR, n = 601) of adenocarcinoma. We varied the cost of false positives (FP), false negatives (FN), and abstained notes and measured total misclassification cost. RESULTS The depression selective classifiers abstained on anywhere from 0% to 97% of notes, and the change in total misclassification cost ranged from -58% to 9%. Selective classifiers abstained on 5%-43% of notes across the GBM and colorectal cancer models. The GBM selective classifier abstained on 43% of notes, which led to improvements in sensitivity (0.94 to 0.96), specificity (0.79 to 0.96), PPV (0.89 to 0.98), and NPV (0.88 to 0.91) when compared to a non-selective classifier and when compared to structured proxy variables. DISCUSSION We showed that selective classifiers outperformed both non-selective classifiers and structured proxy variables for extracting data from unstructured clinical notes. CONCLUSION Selective prediction should be considered when abstaining is preferable to making an incorrect prediction.
Collapse
Affiliation(s)
- Akshay Swaminathan
- Stanford University School of Medicine, Stanford, CA, United States
- Cerebral Inc. Claymont, DE, United States
| | - Ivan Lopez
- Stanford University School of Medicine, Stanford, CA, United States
- Cerebral Inc. Claymont, DE, United States
| | - William Wang
- Department of Biology, Stanford University, Stanford, CA, United States
- Department of Bioengineering, Stanford University, Stanford, CA, United States
| | - Ujwal Srivastava
- Department of Computer Science, Stanford University, Stanford, CA, United States
| | - Edward Tran
- Department of Computer Science, Stanford University, Stanford, CA, United States
- Department of Management Science and Engineering, Stanford University, Stanford, CA, United States
| | | | - Janet Y Wu
- Stanford University School of Medicine, Stanford, CA, United States
| | - Alexander L Ren
- Stanford University School of Medicine, Stanford, CA, United States
| | - Kaitlin Caoili
- Stanford University School of Medicine, Stanford, CA, United States
| | - Brandon Bui
- Department of Human Biology, Stanford University, Stanford, CA, United States
| | - Layth Alkhani
- Department of Bioengineering, Stanford University, Stanford, CA, United States
- Department of Chemistry, Stanford University, Stanford, CA, United States
| | - Susan Lee
- Department of Computer Science, Stanford University, Stanford, CA, United States
| | - Nathan Mohit
- Department of Computer Science, Stanford University, Stanford, CA, United States
- Department of Human Biology, Stanford University, Stanford, CA, United States
| | - Noel Seo
- Department of Sociology, Stanford University, Stanford, CA, United States
| | - Nicholas Macedo
- Department of Biology, Stanford University, Stanford, CA, United States
- Department of Radiology, Stanford University School of Medicine, Stanford, CA, United States
| | - Winson Cheng
- Department of Computer Science, Stanford University, Stanford, CA, United States
- Department of Chemistry, Stanford University, Stanford, CA, United States
| | - Charles Liu
- Department of Surgery, Stanford University School of Medicine, Stanford, CA, United States
| | - Reena Thomas
- Department of Neurology and Neurological Sciences, Stanford Health Care, Stanford, CA, United States
| | - Jonathan H Chen
- Stanford Center for Biomedical Informatics Research, Stanford, CA, United States
- Division of Hospital Medicine, Stanford, CA, United States
- Clinical Excellence Research Center, Stanford, CA, United States
- Department of Medicine, Stanford, CA, United States
| | - Olivier Gevaert
- Stanford Center for Biomedical Informatics Research, Stanford, CA, United States
- Department of Medicine, Stanford, CA, United States
| |
Collapse
|
2
|
Ahlquist KD, Sugden LA, Ramachandran S. Enabling interpretable machine learning for biological data with reliability scores. PLoS Comput Biol 2023; 19:e1011175. [PMID: 37235578 PMCID: PMC10249903 DOI: 10.1371/journal.pcbi.1011175] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2022] [Revised: 06/08/2023] [Accepted: 05/10/2023] [Indexed: 05/28/2023] Open
Abstract
Machine learning tools have proven useful across biological disciplines, allowing researchers to draw conclusions from large datasets, and opening up new opportunities for interpreting complex and heterogeneous biological data. Alongside the rapid growth of machine learning, there have also been growing pains: some models that appear to perform well have later been revealed to rely on features of the data that are artifactual or biased; this feeds into the general criticism that machine learning models are designed to optimize model performance over the creation of new biological insights. A natural question arises: how do we develop machine learning models that are inherently interpretable or explainable? In this manuscript, we describe the SWIF(r) reliability score (SRS), a method building on the SWIF(r) generative framework that reflects the trustworthiness of the classification of a specific instance. The concept of the reliability score has the potential to generalize to other machine learning methods. We demonstrate the utility of the SRS when faced with common challenges in machine learning including: 1) an unknown class present in testing data that was not present in training data, 2) systemic mismatch between training and testing data, and 3) instances of testing data that have missing values for some attributes. We explore these applications of the SRS using a range of biological datasets, from agricultural data on seed morphology, to 22 quantitative traits in the UK Biobank, and population genetic simulations and 1000 Genomes Project data. With each of these examples, we demonstrate how the SRS can allow researchers to interrogate their data and training approach thoroughly, and to pair their domain-specific knowledge with powerful machine-learning frameworks. We also compare the SRS to related tools for outlier and novelty detection, and find that it has comparable performance, with the advantage of being able to operate when some data are missing. The SRS, and the broader discussion of interpretable scientific machine learning, will aid researchers in the biological machine learning space as they seek to harness the power of machine learning without sacrificing rigor and biological insight.
Collapse
Affiliation(s)
- K. D. Ahlquist
- Center for Computational Molecular Biology, Brown University, Providence, Rhode Island, United States of America
- Department of Molecular Biology, Cell Biology, and Biochemistry, Brown University, Providence, Rhode Island, United States of America
| | - Lauren A. Sugden
- Department of Mathematics and Computer Science, Duquesne University, Pittsburgh, Pennsylvania, United States of America
| | - Sohini Ramachandran
- Center for Computational Molecular Biology, Brown University, Providence, Rhode Island, United States of America
- Department of Ecology, Evolution and Organismal Biology, Brown University, Providence, Rhode Island, United States of America
- Data Science Initiative, Brown University, Providence, Rhode Island, United States of America
| |
Collapse
|
3
|
Vision for Improving Pregnancy Health: Innovation and the Future of Pregnancy Research. Reprod Sci 2022; 29:2908-2920. [PMID: 35534766 PMCID: PMC9537127 DOI: 10.1007/s43032-022-00951-w] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2022] [Accepted: 04/15/2022] [Indexed: 10/25/2022]
Abstract
Understanding, predicting, and preventing pregnancy disorders have been a major research target. Nonetheless, the lack of progress is illustrated by research results related to preeclampsia and other hypertensive pregnancy disorders. These remain a major cause of maternal and infant mortality worldwide. There is a general consensus that the rate of progress toward understanding pregnancy disorders lags behind progress in other aspects of human health. In this presentation, we advance an explanation for this failure and suggest solutions. We propose that progress has been impeded by narrowly focused research training and limited imagination and innovation, resulting in the failure to think beyond conventional research approaches and analytical strategies. Investigations have been largely limited to hypothesis-generating approaches constrained by attempts to force poorly defined complex disorders into a single "unifying" hypothesis. Future progress could be accelerated by rethinking this approach. We advise taking advantage of innovative approaches that will generate new research strategies for investigating pregnancy abnormalities. Studies should begin before conception, assessing pregnancy longitudinally, before, during, and after pregnancy. Pregnancy disorders should be defined by pathophysiology rather than phenotype, and state of the art agnostic assessment of data should be adopted to generate new ideas. Taking advantage of new approaches mandates emphasizing innovation, inclusion of large datasets, and use of state of the art experimental and analytical techniques. A revolution in understanding pregnancy-associated disorders will depend on networks of scientists who are driven by an intense biological curiosity, a team spirit, and the tools to make new discoveries.
Collapse
|