1
|
Takefuji Y. Beyond XGBoost and SHAP: Unveiling true feature importance. JOURNAL OF HAZARDOUS MATERIALS 2025; 488:137382. [PMID: 39879771 DOI: 10.1016/j.jhazmat.2025.137382] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/05/2024] [Revised: 01/16/2025] [Accepted: 01/24/2025] [Indexed: 01/31/2025]
Abstract
This paper outlines key machine learning principles, focusing on the use of XGBoost and SHAP values to assist researchers in avoiding analytical pitfalls. XGBoost builds models by incrementally adding decision trees, each addressing the errors of the previous one, which can result in inflated feature importance scores due to the method's emphasis on misclassified examples. While SHAP values provide a theoretically robust way to interpret predictions, their dependence on model structure and feature interactions can introduce biases. The lack of ground truth values complicates model evaluation, as biased feature importance can obscure real relationships with target variables. Ground truth values, representing the actual labels used in model training and validation, are crucial for improving predictive accuracy, serving as benchmarks for comparing model outcomes to true results. However, they do not ensure real associations between features and targets. Instead, they help gauge the model's effectiveness in achieving high accuracy. This paper underscores the necessity for researchers to recognize biases in feature importance and model evaluation, advocating for the use of rigorous statistical methods to enhance the reliability of analyses in machine learning research.
Collapse
Affiliation(s)
- Yoshiyasu Takefuji
- Faculty of Data Science, Musashino University, 3-3-3 Ariake Koto-ku, Tokyo 135-8181, Japan.
| |
Collapse
|
2
|
Colacci M, Huang YQ, Postill G, Zhelnov P, Fennelly O, Verma A, Straus S, Tricco AC. Sociodemographic bias in clinical machine learning models: a scoping review of algorithmic bias instances and mechanisms. J Clin Epidemiol 2025; 178:111606. [PMID: 39532254 DOI: 10.1016/j.jclinepi.2024.111606] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2024] [Revised: 10/22/2024] [Accepted: 11/06/2024] [Indexed: 11/16/2024]
Abstract
BACKGROUND AND OBJECTIVES Clinical machine learning (ML) technologies can sometimes be biased and their use could exacerbate health disparities. The extent to which bias is present, the groups who most frequently experience bias, and the mechanism through which bias is introduced in clinical ML applications is not well described. The objective of this study was to examine instances of bias in clinical ML models. We identified the sociodemographic subgroups PROGRESS that experienced bias and the reported mechanisms of bias introduction. METHODS We searched MEDLINE, EMBASE, PsycINFO, and Web of Science for all studies that evaluated bias on sociodemographic factors within ML algorithms created for the purpose of facilitating clinical care. The scoping review was conducted according to the Joanna Briggs Institute guide and reported using the PRISMA (Preferred Reporting Items for Systematic reviews and Meta-Analyses) extension for scoping reviews. RESULTS We identified 6448 articles, of which 760 reported on a clinical ML model and 91 (12.0%) completed a bias evaluation and met all inclusion criteria. Most studies evaluated a single sociodemographic factor (n = 56, 61.5%). The most frequently evaluated sociodemographic factor was race (n = 59, 64.8%), followed by sex/gender (n = 41, 45.1%), and age (n = 24, 26.4%), with one study (1.1%) evaluating intersectional factors. Of all studies, 74.7% (n = 68) reported that bias was present, 18.7% (n = 17) reported bias was not present, and 6.6% (n = 6) did not state whether bias was present. When present, 87% of studies reported bias against groups with socioeconomic disadvantage. CONCLUSION Most ML algorithms that were evaluated for bias demonstrated bias on sociodemographic factors. Furthermore, most bias evaluations concentrated on race, sex/gender, and age, while other sociodemographic factors and their intersection were infrequently assessed. Given potential health equity implications, bias assessments should be completed for all clinical ML models.
Collapse
Affiliation(s)
- Michael Colacci
- St. Michael's Hospital, Unity Health Toronto, Toronto, Canada; Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Canada.
| | - Yu Qing Huang
- St. Michael's Hospital, Unity Health Toronto, Toronto, Canada; Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Canada
| | - Gemma Postill
- Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Canada; Temerty Faculty of Medicine, University of Toronto, Toronto, Canada
| | - Pavel Zhelnov
- St. Michael's Hospital, Unity Health Toronto, Toronto, Canada
| | - Orna Fennelly
- St. Michael's Hospital, Unity Health Toronto, Toronto, Canada
| | - Amol Verma
- St. Michael's Hospital, Unity Health Toronto, Toronto, Canada; Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Canada; Temerty Faculty of Medicine, University of Toronto, Toronto, Canada
| | - Sharon Straus
- St. Michael's Hospital, Unity Health Toronto, Toronto, Canada; Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Canada; Temerty Faculty of Medicine, University of Toronto, Toronto, Canada
| | - Andrea C Tricco
- St. Michael's Hospital, Unity Health Toronto, Toronto, Canada; Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Canada
| |
Collapse
|
3
|
Schindler L, Beelich H, Röll S, Katsari E, Stracke S, Waltemath D. Applicability of Retrospective and Prospective Gender Scores for Clinical and Health Data: Protocol for a Scoping Review. JMIR Res Protoc 2025; 14:e57669. [PMID: 39832171 PMCID: PMC11791438 DOI: 10.2196/57669] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2024] [Revised: 08/01/2024] [Accepted: 09/18/2024] [Indexed: 01/22/2025] Open
Abstract
BACKGROUND Gender is known to have a strong influence on human health and disease. Despite its relevance to treatment and outcome, gender is insufficiently considered in current health research. One hindering factor is the poor representation of gender information in clinical and health (meta) data. OBJECTIVE We aim to conduct a scoping review of the literature describing gender scores. The review will provide insights into the current application of gender scores in clinical and health settings. The protocol describes how relevant literature will be identified and how gender scores will be evaluated concerning applicability and usability in scientific investigations. METHODS Our scoping review follows the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines. A title and abstract screening was conducted on PubMed, followed by a full-text screening. The inclusion and exclusion criteria were discussed by a team of 5 domain experts, and a data-charting form was developed. The charted data will be categorized, summarized, and analyzed based on the research questions during the scoping review. RESULTS We will report our research results according to the PRISMA-ScR guidelines. The literature retrieval was carried out on June 13, 2024, and resulted in 1202 matches. As of July 2024, the scoping review is in the data extraction phase and we expect to complete and publish the results in the first quarter of 2025. CONCLUSIONS The scoping review lays the foundation for a retrospective gender assessment by identifying scores that can be applied to existing large-scale datasets. Moreover, it will help to formulate recommendations for standardized gender scores in future investigations. INTERNATIONAL REGISTERED REPORT IDENTIFIER (IRRID) DERR1-10.2196/57669.
Collapse
Affiliation(s)
- Lea Schindler
- Medical Informatics Laboratory, University Medicine Greifswald, Greifswald, Germany
| | - Hilke Beelich
- Medical Informatics Laboratory, University Medicine Greifswald, Greifswald, Germany
| | - Selina Röll
- Heart Surgery, University Medicine Greifswald, Greifswald, Germany
| | - Elpiniki Katsari
- Heart Surgery, University Medicine Greifswald, Greifswald, Germany
| | - Sylvia Stracke
- Internal Medicine, University Medicine Greifswald, Greifswald, Germany
| | - Dagmar Waltemath
- Medical Informatics Laboratory, University Medicine Greifswald, Greifswald, Germany
- Core Unit Data Integration Center, University Medicine Greifswald, Greifswald, Germany
| |
Collapse
|
4
|
Kim YE, Serpedin A, Periyakoil P, German D, Rameau A. Sociodemographic reporting in videomics research: a review of practices in otolaryngology - head and neck surgery. Eur Arch Otorhinolaryngol 2024; 281:6047-6056. [PMID: 38704768 DOI: 10.1007/s00405-024-08659-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2024] [Accepted: 04/02/2024] [Indexed: 05/07/2024]
Abstract
OBJECTIVE To assess reporting practices of sociodemographic data in Upper Aerodigestive Tract (UAT) videomics research in Otolaryngology-Head and Neck Surgery (OHNS). STUDY DESIGN Narrative review. METHODS Four online research databases were searched for peer-reviewed articles on videomics and UAT endoscopy in OHNS, published since January 1, 2017. Title and abstract search, followed by a full-text screening was performed. Dataset audit criteria were determined by the MINIMAR reporting standards for patient demographic characteristics, in addition to gender and author affiliations. RESULTS Of the 57 studies that were included, 37% reported any sociodemographic information on their dataset. Among these studies, all reported age, most reported sex (86%), two (10%) reported race, and one (5%) reported ethnicity and socioeconomic status. No studies reported gender. Most studies (84%) included at least one female author, and more than half of the studies (53%) had female first/senior authors, with no significant differences in the rate of sociodemographic reporting in studies with and without female authors (any female author: p = 0.2664; first/senior female author: p > 0.9999). Most studies based in the US reported at least one sociodemographic variable (79%), compared to those in Europe (24%) and in Asia (20%) (p = 0.0012). The rates of sociodemographic reporting in journals of different categories were as follows: clinical OHNS: 44%, clinical non-OHNS: 40%, technical: 42%, interdisciplinary: 10%. CONCLUSIONS There is prevalent underreporting of sociodemographic information in OHNS videomics research utilizing UAT endoscopy. Routine reporting of sociodemographic information should be implemented for AI-based research to help minimize algorithmic biases that have been previously demonstrated. LEVEL OF EVIDENCE: 4
Collapse
Affiliation(s)
- Yeo Eun Kim
- Department of Otolaryngology-Head and Neck Surgery, Weill Cornell Medicine, Sean Parker Institute for the Voice, 240 East 59th St, New York, NY, 10022, USA
| | - Aisha Serpedin
- Department of Otolaryngology-Head and Neck Surgery, Weill Cornell Medicine, Sean Parker Institute for the Voice, 240 East 59th St, New York, NY, 10022, USA
| | - Preethi Periyakoil
- Department of Otolaryngology-Head and Neck Surgery, Weill Cornell Medicine, Sean Parker Institute for the Voice, 240 East 59th St, New York, NY, 10022, USA
| | - Daniel German
- Department of Otolaryngology-Head and Neck Surgery, Weill Cornell Medicine, Sean Parker Institute for the Voice, 240 East 59th St, New York, NY, 10022, USA
| | - Anaïs Rameau
- Department of Otolaryngology-Head and Neck Surgery, Weill Cornell Medicine, Sean Parker Institute for the Voice, 240 East 59th St, New York, NY, 10022, USA.
| |
Collapse
|
5
|
Fabijan A, Zawadzka-Fabijan A, Fabijan R, Zakrzewski K, Nowosławska E, Polis B. Assessing the Accuracy of Artificial Intelligence Models in Scoliosis Classification and Suggested Therapeutic Approaches. J Clin Med 2024; 13:4013. [PMID: 39064053 PMCID: PMC11278075 DOI: 10.3390/jcm13144013] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Revised: 06/30/2024] [Accepted: 07/06/2024] [Indexed: 07/28/2024] Open
Abstract
Background: Open-source artificial intelligence models (OSAIMs) are increasingly being applied in various fields, including IT and medicine, offering promising solutions for diagnostic and therapeutic interventions. In response to the growing interest in AI for clinical diagnostics, we evaluated several OSAIMs-such as ChatGPT 4, Microsoft Copilot, Gemini, PopAi, You Chat, Claude, and the specialized PMC-LLaMA 13B-assessing their abilities to classify scoliosis severity and recommend treatments based on radiological descriptions from AP radiographs. Methods: Our study employed a two-stage methodology, where descriptions of single-curve scoliosis were analyzed by AI models following their evaluation by two independent neurosurgeons. Statistical analysis involved the Shapiro-Wilk test for normality, with non-normal distributions described using medians and interquartile ranges. Inter-rater reliability was assessed using Fleiss' kappa, and performance metrics, like accuracy, sensitivity, specificity, and F1 scores, were used to evaluate the AI systems' classification accuracy. Results: The analysis indicated that although some AI systems, like ChatGPT 4, Copilot, and PopAi, accurately reflected the recommended Cobb angle ranges for disease severity and treatment, others, such as Gemini and Claude, required further calibration. Particularly, PMC-LLaMA 13B expanded the classification range for moderate scoliosis, potentially influencing clinical decisions and delaying interventions. Conclusions: These findings highlight the need for the continuous refinement of AI models to enhance their clinical applicability.
Collapse
Affiliation(s)
- Artur Fabijan
- Department of Neurosurgery, Polish-Mother’s Memorial Hospital Research Institute, 93-338 Lodz, Poland; (K.Z.); (E.N.); (B.P.)
| | - Agnieszka Zawadzka-Fabijan
- Department of Rehabilitation Medicine, Faculty of Health Sciences, Medical University of Lodz, 90-419 Lodz, Poland;
| | | | - Krzysztof Zakrzewski
- Department of Neurosurgery, Polish-Mother’s Memorial Hospital Research Institute, 93-338 Lodz, Poland; (K.Z.); (E.N.); (B.P.)
| | - Emilia Nowosławska
- Department of Neurosurgery, Polish-Mother’s Memorial Hospital Research Institute, 93-338 Lodz, Poland; (K.Z.); (E.N.); (B.P.)
| | - Bartosz Polis
- Department of Neurosurgery, Polish-Mother’s Memorial Hospital Research Institute, 93-338 Lodz, Poland; (K.Z.); (E.N.); (B.P.)
| |
Collapse
|
6
|
Makarov V, Chabbert C, Koletou E, Psomopoulos F, Kurbatova N, Ramirez S, Nelson C, Natarajan P, Neupane B. Good machine learning practices: Learnings from the modern pharmaceutical discovery enterprise. Comput Biol Med 2024; 177:108632. [PMID: 38788373 DOI: 10.1016/j.compbiomed.2024.108632] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2024] [Revised: 05/07/2024] [Accepted: 05/18/2024] [Indexed: 05/26/2024]
Abstract
Machine Learning (ML) and Artificial Intelligence (AI) have become an integral part of the drug discovery and development value chain. Many teams in the pharmaceutical industry nevertheless report the challenges associated with the timely, cost effective and meaningful delivery of ML and AI powered solutions for their scientists. We sought to better understand what these challenges were and how to overcome them by performing an industry wide assessment of the practices in AI and Machine Learning. Here we report results of the systematic business analysis of the personas in the modern pharmaceutical discovery enterprise in relation to their work with the AI and ML technologies. We identify 23 common business problems that individuals in these roles face when they encounter AI and ML technologies at work, and describe best practices (Good Machine Learning Practices) that address these issues.
Collapse
Affiliation(s)
- Vladimir Makarov
- The Pistoia Alliance, 401 Edgewater Place, Suite 600, Wakefield, MA, 01880, USA.
| | | | | | | | | | | | | | | | | |
Collapse
|
7
|
Lin S, Pandit S, Tritsch T, Levy A, Shoja MM. What Goes In, Must Come Out: Generative Artificial Intelligence Does Not Present Algorithmic Bias Across Race and Gender in Medical Residency Specialties. Cureus 2024; 16:e54448. [PMID: 38510858 PMCID: PMC10951939 DOI: 10.7759/cureus.54448] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2023] [Accepted: 02/18/2024] [Indexed: 03/22/2024] Open
Abstract
Objective Artificial Intelligence (AI) has made significant inroads into various domains, including medicine, raising concerns about algorithmic bias. This study investigates the presence of biases in generative AI programs, with a specific focus on gender and racial representations across 19 medical residency specialties. Methodology This comparative study utilized DALL-E2 to generate faces representing 19 distinct residency training specialties, as identified by the Association of American Medical Colleges (AAMC), which were then compared to the AAMC's residency specialty breakdown with respect to race and gender. Results Our findings reveal an alignment between OpenAI's DALL-E2's predictions and the current demographic landscape of medical residents, suggesting an absence of algorithmic bias in this AI model. Conclusion This revelation gives rise to important ethical considerations. While AI excels at pattern recognition, it inherits and mirrors the biases present in its training data. To combat AI bias, addressing real-world disparities is imperative. Initiatives to promote inclusivity and diversity within medicine are commendable and contribute to reshaping medical education. This study underscores the need for ongoing efforts to dismantle barriers and foster inclusivity in historically male-dominated medical fields, particularly for underrepresented populations. Ultimately, our findings underscore the crucial role of real-world data quality in mitigating AI bias. As AI continues to shape healthcare and education, the pursuit of equitable, unbiased AI applications should remain at the forefront of these transformative endeavors.
Collapse
Affiliation(s)
- Shu Lin
- Department of Medical Education, Nova Southeastern University Dr. Kiran C. Patel College of Allopathic Medicine, Fort Lauderdale, USA
| | - Saket Pandit
- Department of Medical Education, Nova Southeastern University Dr. Kiran C. Patel College of Allopathic Medicine, Fort Lauderdale, USA
| | - Tara Tritsch
- Department of Medical Education, Nova Southeastern University Dr. Kiran C. Patel College of Allopathic Medicine, Fort Lauderdale, USA
| | - Arkene Levy
- Department of Medical Education, Nova Southeastern University Dr. Kiran C. Patel College of Allopathic Medicine, Fort Lauderdale, USA
| | - Mohammadali M Shoja
- Department of Medical Education, Nova Southeastern University Dr. Kiran C. Patel College of Allopathic Medicine, Fort Lauderdale, USA
| |
Collapse
|
8
|
Lego VD. Uncovering the gender health data gap. CAD SAUDE PUBLICA 2023; 39:e00065423. [PMID: 37585901 PMCID: PMC10494683 DOI: 10.1590/0102-311xen065423] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2023] [Revised: 06/20/2023] [Accepted: 06/23/2023] [Indexed: 08/18/2023] Open
Affiliation(s)
- Vanessa di Lego
- Vienna Institute of Demography, Austrian Academy of Sciences, Vienna, Austria
| |
Collapse
|
9
|
Jang W, Choi YS, Kim JY, Yon DK, Lee YJ, Chung SH, Kim CY, Yeo SG, Lee J. Artificial Intelligence-Driven Respiratory Distress Syndrome Prediction for Very Low Birth Weight Infants: Korean Multicenter Prospective Cohort Study. J Med Internet Res 2023; 25:e47612. [PMID: 37428525 PMCID: PMC10366668 DOI: 10.2196/47612] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2023] [Revised: 05/04/2023] [Accepted: 06/14/2023] [Indexed: 07/11/2023] Open
Abstract
BACKGROUND Respiratory distress syndrome (RDS) is a disease that commonly affects premature infants whose lungs are not fully developed. RDS results from a lack of surfactant in the lungs. The more premature the infant is, the greater is the likelihood of having RDS. However, even though not all premature infants have RDS, preemptive treatment with artificial pulmonary surfactant is administered in most cases. OBJECTIVE We aimed to develop an artificial intelligence model to predict RDS in premature infants to avoid unnecessary treatment. METHODS In this study, 13,087 very low birth weight infants who were newborns weighing less than 1500 grams were assessed in 76 hospitals of the Korean Neonatal Network. To predict RDS in very low birth weight infants, we used basic infant information, maternity history, pregnancy/birth process, family history, resuscitation procedure, and test results at birth such as blood gas analysis and Apgar score. The prediction performances of 7 different machine learning models were compared, and a 5-layer deep neural network was proposed in order to enhance the prediction performance from the selected features. An ensemble approach combining multiple models from the 5-fold cross-validation was subsequently developed. RESULTS Our proposed ensemble 5-layer deep neural network consisting of the top 20 features provided high sensitivity (83.03%), specificity (87.50%), accuracy (84.07%), balanced accuracy (85.26%), and area under the curve (0.9187). Based on the model that we developed, a public web application that enables easy access for the prediction of RDS in premature infants was deployed. CONCLUSIONS Our artificial intelligence model may be useful for preparations for neonatal resuscitation, particularly in cases involving the delivery of very low birth weight infants, as it can aid in predicting the likelihood of RDS and inform decisions regarding the administration of surfactant.
Collapse
Affiliation(s)
- Woocheol Jang
- Biomedical Engineering, Kyung Hee University, Yongin-si, Republic of Korea
| | - Yong Sung Choi
- Department of Pediatrics, Kyung Hee University Medical Center, Kyung Hee University College of Medicine, Seoul, Republic of Korea
| | - Ji Yoo Kim
- Department of Pediatrics, Kyung Hee University Medical Center, Kyung Hee University College of Medicine, Seoul, Republic of Korea
| | - Dong Keon Yon
- Department of Pediatrics, Kyung Hee University Medical Center, Kyung Hee University College of Medicine, Seoul, Republic of Korea
| | - Young Joo Lee
- Department of Obstetrics and Gynecology, Kyung Hee University Medical Center, Kyung Hee University College of Medicine, Seoul, Republic of Korea
| | - Sung-Hoon Chung
- Department of Pediatrics, Kyung Hee University Hospital at Gangdong, Kyung Hee University College of Medicine, Seoul, Republic of Korea
| | - Chae Young Kim
- Department of Pediatrics, Kyung Hee University Hospital at Gangdong, Kyung Hee University College of Medicine, Seoul, Republic of Korea
| | - Seung Geun Yeo
- Department of Otorhinolaryngology Head and Neck Surgery, Kyung Hee University Medical Center, Kyung Hee University School of Medicine, Seoul, Republic of Korea
| | - Jinseok Lee
- Biomedical Engineering, Kyung Hee University, Yongin-si, Republic of Korea
| |
Collapse
|
10
|
Kidwai-Khan F, Wang R, Skanderson M, Brandt CA, Fodeh S, Womack JA. A Roadmap to Artificial Intelligence (AI): Methods for Designing and Building AI ready Data for Women's Health Studies. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2023:2023.05.25.23290399. [PMID: 37398113 PMCID: PMC10312839 DOI: 10.1101/2023.05.25.23290399] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/04/2023]
Abstract
Objectives Evaluating methods for building data frameworks for application of AI in large scale datasets for women's health studies. Methods We created methods for transforming raw data to a data framework for applying machine learning (ML) and natural language processing (NLP) techniques for predicting falls and fractures. Results Prediction of falls was higher in women compared to men. Information extracted from radiology reports was converted to a matrix for applying machine learning. For fractures, by applying specialized algorithms, we extracted snippets from dual x-ray absorptiometry (DXA) scans for meaningful terms usable for predicting fracture risk. Discussion Life cycle of data from raw to analytic form includes data governance, cleaning, management, and analysis. For applying AI, data must be prepared optimally to reduce algorithmic bias. Conclusion Algorithmic bias is harmful for research using AI methods. Building AI ready data frameworks that improve efficiency can be especially valuable for women's health.
Collapse
Affiliation(s)
- Farah Kidwai-Khan
- Yale School of Medicine, New Haven, Connecticut, USA
- VA Connecticut Healthcare System, West Haven, Connecticut, USA
| | - Rixin Wang
- Yale School of Medicine, New Haven, Connecticut, USA
- VA Connecticut Healthcare System, West Haven, Connecticut, USA
| | | | - Cynthia A. Brandt
- Yale School of Medicine, New Haven, Connecticut, USA
- VA Connecticut Healthcare System, West Haven, Connecticut, USA
| | - Samah Fodeh
- Yale School of Medicine, New Haven, Connecticut, USA
- VA Connecticut Healthcare System, West Haven, Connecticut, USA
| | - Julie A. Womack
- VA Connecticut Healthcare System, West Haven, Connecticut, USA
- Yale School of Nursing, New Haven, Connecticut, USA
| |
Collapse
|