1
|
Yousefi F, Dehnavieh R, Laberge M, Gagnon MP, Ghaemi MM, Nadali M, Azizi N. Opportunities, challenges, and requirements for Artificial Intelligence (AI) implementation in Primary Health Care (PHC): a systematic review. BMC PRIMARY CARE 2025; 26:196. [PMID: 40490689 DOI: 10.1186/s12875-025-02785-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/03/2024] [Accepted: 03/11/2025] [Indexed: 06/11/2025]
Abstract
BACKGROUND Artificial Intelligence (AI) has significantly reshaped Primary Health Care (PHC), offering various possibilities and complexities across all functional dimensions. The objective is to review and synthesize available evidence on the opportunities, challenges, and requirements of AI implementation in PHC based on the Primary Care Evaluation Tool (PCET). METHODS We conducted a systematic review, following the Cochrane Collaboration method, to identify the latest evidence regarding AI implementation in PHC. A comprehensive search across eight databases- PubMed, Web of Science, Scopus, Science Direct, Embase, CINAHL, IEEE, and Cochrane was conducted using MeSH terms alongside the SPIDER framework to pinpoint quantitative and qualitative literature published from 2000 to 2024. Two reviewers independently applied inclusion and exclusion criteria, guided by the SPIDER framework, to review full texts and extract data. We synthesized extracted data from the study characteristics, opportunities, challenges, and requirements, employing thematic-framework analysis, according to the PCET model. The quality of the studies was evaluated using the JBI critical appraisal tools. RESULTS In this review, we included a total of 109 articles, most of which were conducted in North America (n = 49, 44%), followed by Europe (n = 36, 33%). The included studies employed a diverse range of study designs. Using the PCET model, we categorized AI-related opportunities, challenges, and requirements across four key dimensions. The greatest opportunities for AI integration in PHC were centered on enhancing comprehensive service delivery, particularly by improving diagnostic accuracy, optimizing screening programs, and advancing early disease prediction. However, the most challenges emerged within the stewardship and resource generation functions, with key concerns related to data security and privacy, technical performance issues, and limitations in data accessibility. Ensuring successful AI integration requires a robust stewardship function, strategic investments in resource generation, and a collaborative approach that fosters co-development, scientific advancements, and continuous evaluation. CONCLUSIONS Successful AI integration in PHC requires a coordinated, multidimensional approach, with stewardship, resource generation, and financing playing key roles in enabling service delivery. Addressing existing knowledge gaps, examining interactions among these dimensions, and fostering a collaborative approach in developing AI solutions among stakeholders are essential steps toward achieving an equitable and efficient AI-driven PHC system. PROTOCOL Registered in Open Science Framework (OSF) ( https://doi.org/10.17605/OSF.IO/HG2DV ).
Collapse
Affiliation(s)
- Farzaneh Yousefi
- Department of Management, Policy and Health Economics, Faculty of Medical Information and Management, Candidate in Health Services Management, Kerman University of Medical Sciences, Kerman, Iran
- Faculty of Nursing, Research Professional in Health Services Research, Laval University, Quebec, Canada
| | - Reza Dehnavieh
- Health Foresight and Innovation Research Center, , Institute for Futures Studies in Health, Kerman University of Medical Sciences, Kerman, Iran.
- Department of Management, Policy and Health Economics, Faculty of Medical Information and Management, Professor of Health Services Management, Kerman University of Medical Sciences, Kerman, Iran.
| | - Maude Laberge
- Faculté de Médecine, Université Laval, Quebec, QC, Canada
- Centre de Recherche du CHU de Québec-Université Laval (CRCHUQ), Quebec, QC, Canada
| | - Marie-Pierre Gagnon
- CHU de Québec-Université Laval Research Centre, Québec, Canada
- Faculty of Nursing, Université Laval,, Québec, Canada
| | - Mohammad Mehdi Ghaemi
- Medical Informatics Research Center, Institute for Futures Studies in Health, Kerman University of Medical Sciences, Kerman, Iran
| | - Mohsen Nadali
- Master of Business Administration, Mehralborz University, Tehran, Iran
| | - Najmeh Azizi
- Department of Management, Policy and Health Economics, Faculty of Medical Information and Management, Student in Health Services Management, Kerman University of Medical Sciences, Kerman, Iran
| |
Collapse
|
2
|
Bouguettaya A, Team V, Stuart EM, Aboujaoude E. AI-driven report-generation tools in mental healthcare: A review of commercial tools. Gen Hosp Psychiatry 2025; 94:150-158. [PMID: 40088857 DOI: 10.1016/j.genhosppsych.2025.02.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/10/2024] [Revised: 02/21/2025] [Accepted: 02/21/2025] [Indexed: 03/17/2025]
Abstract
Artificial intelligence (AI) systems are increasingly being integrated in clinical care, including for AI-powered note-writing. We aimed to develop and apply a scale for assessing mental health electronic health records (EHRs) that use large language models (LLMs) for note-writing, focusing on their features, security, and ethics. The assessment involved analyzing product information and directly querying vendors about their systems. On their websites, the majority of vendors provided comprehensive information on data protection, privacy measures, multi-platform availability, patient access features, software update history, and Meaningful Use compliance. Most products clearly indicated the LLM's capabilities in creating customized reports or functioning as a co-pilot. However, critical information was often absent, including details on LLM training methodologies, the specific LLM used, bias correction techniques, and methods for evaluating the evidence base. The lack of transparency regarding LLM specifics and bias mitigation strategies raises concerns about the ethical implementation and reliability of these systems in clinical practice. While LLM-enhanced EHRs show promise in alleviating the documentation burden for mental health professionals, there is a pressing need for greater transparency and standardization in reporting LLM-related information. We propose recommendations for the future development and implementation of these systems to ensure they meet the highest standards of security, ethics, and clinical care.
Collapse
Affiliation(s)
- Ayoub Bouguettaya
- Department of Biomedical Sciences, Cedars-Sinai Medical Center, Los Angeles, CA, United States; School of Nursing and Midwifery, Monash University, Melbourne, Victoria, Australia
| | - Victoria Team
- School of Nursing and Midwifery, Monash University, Melbourne, Victoria, Australia
| | - Elizabeth M Stuart
- Jonathan Jaques Children's Cancer Institute, Miller Children's & Women's Hospital Long Beach, Long Beach, CA, United States
| | - Elias Aboujaoude
- Department of Biomedical Sciences, Cedars-Sinai Medical Center, Los Angeles, CA, United States; Department of Psychiatry and Behavioral Sciences, Stanford University, Stanford, CA, United States.
| |
Collapse
|
3
|
Sung L, Brudno M, Caesar MCW, Verma AA, Buchsbaum B, Retnakaran R, Giannakeas V, Kushki A, Bader GD, Lasthiotakis H, Mamdani M, Strug L. Approaches to identify scenarios for data science implementations within healthcare settings: recommendations based on experiences at multiple academic institutions. Front Digit Health 2025; 7:1511943. [PMID: 40161559 PMCID: PMC11949942 DOI: 10.3389/fdgth.2025.1511943] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2024] [Accepted: 02/27/2025] [Indexed: 04/02/2025] Open
Abstract
Objectives To describe successful and unsuccessful approaches to identify scenarios for data science implementations within healthcare settings and to provide recommendations for future scenario identification procedures. Materials and methods Representatives from seven Toronto academic healthcare institutions participated in a one-day workshop. Each institution was asked to provide an introduction to their clinical data science program and to provide an example of a successful and unsuccessful approach to scenario identification at their institution. Using content analysis, common observations were summarized. Results Observations were coalesced to idea generation and value proposition, prioritization, approval and champions. Successful experiences included promoting a portfolio of ideas, articulating value proposition, ensuring alignment with organization priorities, ensuring approvers can adjudicate feasibility and identifying champions willing to take ownership over the projects. Conclusion Based on academic healthcare data science program experiences, we provided recommendations for approaches to identify scenarios for data science implementations within healthcare settings.
Collapse
Affiliation(s)
- Lillian Sung
- Department of Paediatrics, The Hospital for Sick Children, Institute of Health Policy Management & Evaluation, University of Toronto, Toronto, ON, Canada
| | - Michael Brudno
- Department of Computer Science, Vector Institute for Artificial Intelligence, University Health Network, University of Toronto, Toronto, ON, Canada
| | - Michael C. W. Caesar
- Institute of Health Policy Management & Evaluation, University Health Network, University of Toronto, Toronto, ON, Canada
| | - Amol A. Verma
- Department of Medicine, Department of Laboratory Medicine and Pathobiology, and Institution of Health Policy Management & Evaluation; St. Michael’s Hospital, Unity Health Toronto, University of Toronto, Toronto, ON, Canada
| | - Brad Buchsbaum
- Department of Psychology, Rotman Research Institute, Baycrest Centre, University of Toronto, Toronto, ON, Canada
| | - Ravi Retnakaran
- Division of Endocrinology, Lunenfeld-Tanenbaum Research Institute, Mount Sinai Hospital, Toronto, ON, Canada
| | - Vasily Giannakeas
- Women’s College Research Institute, Women’s College Hospital, Toronto, ON, Canada
| | - Azadeh Kushki
- Institute of Biomedical Engineering, Holland Bloorview Kids Rehabilitation Hospital, University of Toronto, Toronto, ON, Canada
| | - Gary D. Bader
- Department of Molecular Genetics, Temerty Faculty of Medicine, Toronto, ON, Canada
| | | | - Muhammad Mamdani
- Temerty Faculty of Medicine, Centre for Artificial Intelligence Education and Research in Medicine, Unity Health Toronto, University of Toronto, Toronto, ON, Canada
| | - Lisa Strug
- Data Sciences Institute, University of Toronto, Toronto, ON, Canada
| |
Collapse
|
4
|
Kaewboonlert N, Poontananggul J, Pongsuwan N, Bhakdisongkhram G. Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study. JMIR MEDICAL EDUCATION 2025; 11:e58898. [PMID: 39846415 PMCID: PMC11745146 DOI: 10.2196/58898] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Revised: 05/22/2024] [Accepted: 12/04/2024] [Indexed: 01/24/2025]
Abstract
Background Artificial intelligence (AI) has become widely applied across many fields, including medical education. Content validation and its answers are based on training datasets and the optimization of each model. The accuracy of large language model (LLMs) in basic medical examinations and factors related to their accuracy have also been explored. Objective We evaluated factors associated with the accuracy of LLMs (GPT-3.5, GPT-4, Google Bard, and Microsoft Bing) in answering multiple-choice questions from basic medical science examinations. Methods We used questions that were closely aligned with the content and topic distribution of Thailand's Step 1 National Medical Licensing Examination. Variables such as the difficulty index, discrimination index, and question characteristics were collected. These questions were then simultaneously input into ChatGPT (with GPT-3.5 and GPT-4), Microsoft Bing, and Google Bard, and their responses were recorded. The accuracy of these LLMs and the associated factors were analyzed using multivariable logistic regression. This analysis aimed to assess the effect of various factors on model accuracy, with results reported as odds ratios (ORs). Results The study revealed that GPT-4 was the top-performing model, with an overall accuracy of 89.07% (95% CI 84.76%-92.41%), significantly outperforming the others (P<.001). Microsoft Bing followed with an accuracy of 83.69% (95% CI 78.85%-87.80%), GPT-3.5 at 67.02% (95% CI 61.20%-72.48%), and Google Bard at 63.83% (95% CI 57.92%-69.44%). The multivariable logistic regression analysis showed a correlation between question difficulty and model performance, with GPT-4 demonstrating the strongest association. Interestingly, no significant correlation was found between model accuracy and question length, negative wording, clinical scenarios, or the discrimination index for most models, except for Google Bard, which showed varying correlations. Conclusions The GPT-4 and Microsoft Bing models demonstrated equal and superior accuracy compared to GPT-3.5 and Google Bard in the domain of basic medical science. The accuracy of these models was significantly influenced by the item's difficulty index, indicating that the LLMs are more accurate when answering easier questions. This suggests that the more accurate models, such as GPT-4 and Bing, can be valuable tools for understanding and learning basic medical science concepts.
Collapse
Affiliation(s)
- Naritsaret Kaewboonlert
- Institute of Medicine, Suranaree University of Technology, 111 University Avenue, Nakhon Ratchasima, 30000, Thailand, 66 44223956
| | - Jiraphon Poontananggul
- Institute of Medicine, Suranaree University of Technology, 111 University Avenue, Nakhon Ratchasima, 30000, Thailand, 66 44223956
| | - Natthipong Pongsuwan
- Institute of Medicine, Suranaree University of Technology, 111 University Avenue, Nakhon Ratchasima, 30000, Thailand, 66 44223956
| | - Gun Bhakdisongkhram
- Institute of Medicine, Suranaree University of Technology, 111 University Avenue, Nakhon Ratchasima, 30000, Thailand, 66 44223956
| |
Collapse
|
5
|
Abou Chaar MK, Grigsby-Rocca G, Huang M, Blackmon SH. ChatGPT vs Expert-Guided Care Pathways for Postesophagectomy Symptom Management. ANNALS OF THORACIC SURGERY SHORT REPORTS 2024; 2:674-679. [PMID: 39790627 PMCID: PMC11708366 DOI: 10.1016/j.atssr.2024.06.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Accepted: 06/04/2024] [Indexed: 01/12/2025]
Abstract
Background The objective of this study was to compare generative artificial intelligence-initiated care pathways, using ChatGPT, with expert-guided consensus-initiated care pathways from AskMayoExpert (AME) for symptom management of esophageal cancer patients after esophagectomy. Methods A formal protocol for development of 9 AME care pathways was followed for specific patient-identified domains after esophagectomy for esophageal cancer. Domain scores were measured and assessed through the Upper Digestive Disease tool. These care pathways were developed by experts validated by a consensus-driven methodology. ChatGPT was used to answer specific questions similar to the AME care pathway on April 9, 2023, and March 28, 2024. To compare outcomes, answers were recorded, and algorithms were compared with a survey tool composed of 5 questions. Results Both modalities were able to provide a clear definition with multidirectional management options for all 9 domains: dysphagia, generalized dumping, gastrointestinal dumping, pain, regurgitation, heartburn, nausea, physical health, and mental health. When provided with a simple prompt, ChatGPT 3.5 failed to provide a comprehensive stepwise approach for providers, any testing recommendations, or any form of triage process. However, ChatGPT 4.0 provided plans, similar to AME care pathways, when a sophisticated prompt was used. Conclusions Generative artificial intelligence-initiated care pathways can be used by physicians as a supplementary tool to guide provider management of patients with complex symptoms after esophagectomy. This technology will continue to advance but is currently insufficient to solely guide clinical management of complex patients with severe symptoms.
Collapse
Affiliation(s)
- Mohamad K. Abou Chaar
- Division of Thoracic Surgery, Department of Surgery, Mayo Clinic, Rochester, Minnesota
| | | | - Ming Huang
- Department of Health Data Science and Artificial Intelligence, University of Texas Health Science Center at Houston, Houston, Texas
| | | |
Collapse
|
6
|
Collins BX, Bélisle-Pipon JC, Evans BJ, Ferryman K, Jiang X, Nebeker C, Novak L, Roberts K, Were M, Yin Z, Ravitsky V, Coco J, Hendricks-Sturrup R, Williams I, Clayton EW, Malin BA, Bridge2AI Ethics and Trustworthy AI Working Group. Addressing ethical issues in healthcare artificial intelligence using a lifecycle-informed process. JAMIA Open 2024; 7:ooae108. [PMID: 39553826 PMCID: PMC11565898 DOI: 10.1093/jamiaopen/ooae108] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2024] [Revised: 08/19/2024] [Accepted: 10/04/2024] [Indexed: 11/19/2024] Open
Abstract
Objectives Artificial intelligence (AI) proceeds through an iterative and evaluative process of development, use, and refinement which may be characterized as a lifecycle. Within this context, stakeholders can vary in their interests and perceptions of the ethical issues associated with this rapidly evolving technology in ways that can fail to identify and avert adverse outcomes. Identifying issues throughout the AI lifecycle in a systematic manner can facilitate better-informed ethical deliberation. Materials and Methods We analyzed existing lifecycles from within the current literature for ethical issues of AI in healthcare to identify themes, which we relied upon to create a lifecycle that consolidates these themes into a more comprehensive lifecycle. We then considered the potential benefits and harms of AI through this lifecycle to identify ethical questions that can arise at each step and to identify where conflicts and errors could arise in ethical analysis. We illustrated the approach in 3 case studies that highlight how different ethical dilemmas arise at different points in the lifecycle. Results Discussion and Conclusion Through case studies, we show how a systematic lifecycle-informed approach to the ethical analysis of AI enables mapping of the effects of AI onto different steps to guide deliberations on benefits and harms. The lifecycle-informed approach has broad applicability to different stakeholders and can facilitate communication on ethical issues for patients, healthcare professionals, research participants, and other stakeholders.
Collapse
Affiliation(s)
- Benjamin X Collins
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37203, United States
- Center for Biomedical Ethics and Society, Vanderbilt University Medical Center, Nashville, TN 37203, United States
- Department of Medicine, Vanderbilt University Medical Center, Nashville, TN 37232, United States
| | | | - Barbara J Evans
- Levin College of Law, University of Florida, Gainesville, FL 32611, United States
- Herbert Wertheim College of Engineering, University of Florida, Gainesville, FL 32611, United States
| | - Kadija Ferryman
- Berman Institute of Bioethics, Johns Hopkins University, Baltimore, MD 21205, United States
| | - Xiaoqian Jiang
- McWilliams School of Biomedical Informatics, UTHealth Houston, Houston, TX 77030, United States
| | - Camille Nebeker
- Herbert Wertheim School of Public Health and Human Longevity Science, University of California, San Diego, La Jolla, CA 92093, United States
| | - Laurie Novak
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37203, United States
| | - Kirk Roberts
- McWilliams School of Biomedical Informatics, UTHealth Houston, Houston, TX 77030, United States
| | - Martin Were
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37203, United States
- Department of Medicine, Vanderbilt University Medical Center, Nashville, TN 37232, United States
| | - Zhijun Yin
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37203, United States
- Department of Computer Science, Vanderbilt University, Nashville, TN 37212, United States
| | | | - Joseph Coco
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37203, United States
| | - Rachele Hendricks-Sturrup
- National Alliance against Disparities in Patient Health, Woodbridge, VA 22191, United States
- Margolis Center for Health Policy, Duke University, Washington, DC 20004, United States
| | - Ishan Williams
- School of Nursing, University of Virginia, Charlottesville, VA 22903, United States
| | - Ellen W Clayton
- Center for Biomedical Ethics and Society, Vanderbilt University Medical Center, Nashville, TN 37203, United States
- Law School, Vanderbilt University, Nashville, TN 37203, United States
| | - Bradley A Malin
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN 37203, United States
- Department of Computer Science, Vanderbilt University, Nashville, TN 37212, United States
| | | |
Collapse
|
7
|
Sideris K, Weir CR, Schmalfuss C, Hanson H, Pipke M, Tseng PH, Lewis N, Sallam K, Bozkurt B, Hanff T, Schofield R, Larimer K, Kyriakopoulos CP, Taleb I, Brinker L, Curry T, Knecht C, Butler JM, Stehlik J. Artificial intelligence predictive analytics in heart failure: results of the pilot phase of a pragmatic randomized clinical trial. J Am Med Inform Assoc 2024; 31:919-928. [PMID: 38341800 PMCID: PMC10990545 DOI: 10.1093/jamia/ocae017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Revised: 12/20/2023] [Accepted: 01/17/2024] [Indexed: 02/13/2024] Open
Abstract
OBJECTIVES We conducted an implementation planning process during the pilot phase of a pragmatic trial, which tests an intervention guided by artificial intelligence (AI) analytics sourced from noninvasive monitoring data in heart failure patients (LINK-HF2). MATERIALS AND METHODS A mixed-method analysis was conducted at 2 pilot sites. Interviews were conducted with 12 of 27 enrolled patients and with 13 participating clinicians. iPARIHS constructs were used for interview construction to identify workflow, communication patterns, and clinician's beliefs. Interviews were transcribed and analyzed using inductive coding protocols to identify key themes. Behavioral response data from the AI-generated notifications were collected. RESULTS Clinicians responded to notifications within 24 hours in 95% of instances, with 26.7% resulting in clinical action. Four implementation themes emerged: (1) High anticipatory expectations for reliable patient communications, reduced patient burden, and less proactive provider monitoring. (2) The AI notifications required a differential and tailored balance of trust and action advice related to role. (3) Clinic experience with other home-based programs influenced utilization. (4) Responding to notifications involved significant effort, including electronic health record (EHR) review, patient contact, and consultation with other clinicians. DISCUSSION Clinician's use of AI data is a function of beliefs regarding the trustworthiness and usefulness of the data, the degree of autonomy in professional roles, and the cognitive effort involved. CONCLUSION The implementation planning analysis guided development of strategies that addressed communication technology, patient education, and EHR integration to reduce clinician and patient burden in the subsequent main randomized phase of the trial. Our results provide important insights into the unique implications of implementing AI analytics into clinical workflow.
Collapse
Affiliation(s)
- Konstantinos Sideris
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States
| | - Charlene R Weir
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Department of Biomedical Informatics, School of Medicine, University of Utah, Salt Lake City, UT 84108, United States
| | - Carsten Schmalfuss
- Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States
| | - Heather Hanson
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States
| | - Matt Pipke
- PhysIQ, Inc., Chicago, IL 60563, United States
| | - Po-He Tseng
- PhysIQ, Inc., Chicago, IL 60563, United States
| | - Neil Lewis
- Cardiology Section, Medical Service, Hunter Holmes McGuire Veterans Medical Center, Richmond, VA 23249, United States
- Department of Internal Medicine, Division of Cardiovascular Disease, Virginia Commonwealth University, Richmond, VA 23249, United States
| | - Karim Sallam
- Cardiology Section, Medical Service, VA Palo Alto Health Care System, Palo Alto, CA 94304, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, Stanford University School of Medicine, Stanford, CA 94305, United States
| | - Biykem Bozkurt
- Cardiology Section, Medical Service, Michael E. DeBakey VA Medical Center, Houston, TX 77030, United States
- Section of Cardiology, Department of Medicine, Baylor College of Medicine, Houston, TX 77030, United States
| | - Thomas Hanff
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States
| | - Richard Schofield
- Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States
| | | | - Christos P Kyriakopoulos
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States
| | - Iosif Taleb
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States
| | - Lina Brinker
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States
| | - Tempa Curry
- Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States
| | - Cheri Knecht
- Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States
| | - Jorie M Butler
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Department of Biomedical Informatics, School of Medicine, University of Utah, Salt Lake City, UT 84108, United States
| | - Josef Stehlik
- Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States
- Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States
| |
Collapse
|
8
|
Yim D, Khuntia J, Parameswaran V, Meyers A. Preliminary Evidence of the Use of Generative AI in Health Care Clinical Services: Systematic Narrative Review. JMIR Med Inform 2024; 12:e52073. [PMID: 38506918 PMCID: PMC10993141 DOI: 10.2196/52073] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/21/2023] [Revised: 10/12/2023] [Accepted: 01/30/2024] [Indexed: 03/21/2024] Open
Abstract
BACKGROUND Generative artificial intelligence tools and applications (GenAI) are being increasingly used in health care. Physicians, specialists, and other providers have started primarily using GenAI as an aid or tool to gather knowledge, provide information, train, or generate suggestive dialogue between physicians and patients or between physicians and patients' families or friends. However, unless the use of GenAI is oriented to be helpful in clinical service encounters that can improve the accuracy of diagnosis, treatment, and patient outcomes, the expected potential will not be achieved. As adoption continues, it is essential to validate the effectiveness of the infusion of GenAI as an intelligent technology in service encounters to understand the gap in actual clinical service use of GenAI. OBJECTIVE This study synthesizes preliminary evidence on how GenAI assists, guides, and automates clinical service rendering and encounters in health care The review scope was limited to articles published in peer-reviewed medical journals. METHODS We screened and selected 0.38% (161/42,459) of articles published between January 1, 2020, and May 31, 2023, identified from PubMed. We followed the protocols outlined in the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines to select highly relevant studies with at least 1 element on clinical use, evaluation, and validation to provide evidence of GenAI use in clinical services. The articles were classified based on their relevance to clinical service functions or activities using the descriptive and analytical information presented in the articles. RESULTS Of 161 articles, 141 (87.6%) reported using GenAI to assist services through knowledge access, collation, and filtering. GenAI was used for disease detection (19/161, 11.8%), diagnosis (14/161, 8.7%), and screening processes (12/161, 7.5%) in the areas of radiology (17/161, 10.6%), cardiology (12/161, 7.5%), gastrointestinal medicine (4/161, 2.5%), and diabetes (6/161, 3.7%). The literature synthesis in this study suggests that GenAI is mainly used for diagnostic processes, improvement of diagnosis accuracy, and screening and diagnostic purposes using knowledge access. Although this solves the problem of knowledge access and may improve diagnostic accuracy, it is oriented toward higher value creation in health care. CONCLUSIONS GenAI informs rather than assisting or automating clinical service functions in health care. There is potential in clinical service, but it has yet to be actualized for GenAI. More clinical service-level evidence that GenAI is used to streamline some functions or provides more automated help than only information retrieval is needed. To transform health care as purported, more studies related to GenAI applications must automate and guide human-performed services and keep up with the optimism that forward-thinking health care organizations will take advantage of GenAI.
Collapse
Affiliation(s)
- Dobin Yim
- Loyola University, Maryland, MD, United States
| | - Jiban Khuntia
- University of Colorado Denver, Denver, CO, United States
| | | | - Arlen Meyers
- University of Colorado Denver, Denver, CO, United States
| |
Collapse
|
9
|
Reddy S. Generative AI in healthcare: an implementation science informed translational path on application, integration and governance. Implement Sci 2024; 19:27. [PMID: 38491544 PMCID: PMC10941464 DOI: 10.1186/s13012-024-01357-9] [Citation(s) in RCA: 62] [Impact Index Per Article: 62.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Accepted: 03/06/2024] [Indexed: 03/18/2024] Open
Abstract
BACKGROUND Artificial intelligence (AI), particularly generative AI, has emerged as a transformative tool in healthcare, with the potential to revolutionize clinical decision-making and improve health outcomes. Generative AI, capable of generating new data such as text and images, holds promise in enhancing patient care, revolutionizing disease diagnosis and expanding treatment options. However, the utility and impact of generative AI in healthcare remain poorly understood, with concerns around ethical and medico-legal implications, integration into healthcare service delivery and workforce utilisation. Also, there is not a clear pathway to implement and integrate generative AI in healthcare delivery. METHODS This article aims to provide a comprehensive overview of the use of generative AI in healthcare, focusing on the utility of the technology in healthcare and its translational application highlighting the need for careful planning, execution and management of expectations in adopting generative AI in clinical medicine. Key considerations include factors such as data privacy, security and the irreplaceable role of clinicians' expertise. Frameworks like the technology acceptance model (TAM) and the Non-Adoption, Abandonment, Scale-up, Spread and Sustainability (NASSS) model are considered to promote responsible integration. These frameworks allow anticipating and proactively addressing barriers to adoption, facilitating stakeholder participation and responsibly transitioning care systems to harness generative AI's potential. RESULTS Generative AI has the potential to transform healthcare through automated systems, enhanced clinical decision-making and democratization of expertise with diagnostic support tools providing timely, personalized suggestions. Generative AI applications across billing, diagnosis, treatment and research can also make healthcare delivery more efficient, equitable and effective. However, integration of generative AI necessitates meticulous change management and risk mitigation strategies. Technological capabilities alone cannot shift complex care ecosystems overnight; rather, structured adoption programs grounded in implementation science are imperative. CONCLUSIONS It is strongly argued in this article that generative AI can usher in tremendous healthcare progress, if introduced responsibly. Strategic adoption based on implementation science, incremental deployment and balanced messaging around opportunities versus limitations helps promote safe, ethical generative AI integration. Extensive real-world piloting and iteration aligned to clinical priorities should drive development. With conscientious governance centred on human wellbeing over technological novelty, generative AI can enhance accessibility, affordability and quality of care. As these models continue advancing rapidly, ongoing reassessment and transparent communication around their strengths and weaknesses remain vital to restoring trust, realizing positive potential and, most importantly, improving patient outcomes.
Collapse
Affiliation(s)
- Sandeep Reddy
- Deakin School of Medicine, Waurn Ponds, Geelong, VIC, 3215, Australia.
| |
Collapse
|
10
|
Bhatla A, Kim CH, Nimbalkar M, Ng‐Thow‐Hing AS, Isakadze N, Spaulding E, Zaleski A, Craig KJ, Verbrugge DJ, Dunn P, Nag D, Bankar D, Martin SS, Marvel FA. Cardiac Rehabilitation Enabled With Health Technology: Innovative Models of Care Delivery and Policy to Enhance Health Equity. J Am Heart Assoc 2024; 13:e031621. [PMID: 38226509 PMCID: PMC10926793 DOI: 10.1161/jaha.123.031621] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/03/2023] [Accepted: 10/11/2023] [Indexed: 01/17/2024]
Affiliation(s)
- Anjali Bhatla
- Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
| | - Chang H. Kim
- Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Center for Mobile Technologies to Achieve Equity in Cardiovascular Health (mTECH Center), an AHA Health Technology & Innovation SFRN CenterBaltimoreMD
| | - Mansi Nimbalkar
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Center for Mobile Technologies to Achieve Equity in Cardiovascular Health (mTECH Center), an AHA Health Technology & Innovation SFRN CenterBaltimoreMD
| | - Anthony Sky Ng‐Thow‐Hing
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
| | - Nino Isakadze
- Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Center for Mobile Technologies to Achieve Equity in Cardiovascular Health (mTECH Center), an AHA Health Technology & Innovation SFRN CenterBaltimoreMD
| | - Erin Spaulding
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Center for Mobile Technologies to Achieve Equity in Cardiovascular Health (mTECH Center), an AHA Health Technology & Innovation SFRN CenterBaltimoreMD
- School of NursingJohns Hopkins UniversityBaltimoreMD
- Welch Center for Prevention, Epidemiology, and Clinical ResearchJohns Hopkins Bloomberg School of Public HealthBaltimoreMD
| | | | | | | | | | | | | | - Seth S. Martin
- Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Center for Mobile Technologies to Achieve Equity in Cardiovascular Health (mTECH Center), an AHA Health Technology & Innovation SFRN CenterBaltimoreMD
| | - Francoise A. Marvel
- Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Digital Health Innovation Laboratory, Ciccarone Center for the Prevention of Cardiovascular Disease, Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMD
- Center for Mobile Technologies to Achieve Equity in Cardiovascular Health (mTECH Center), an AHA Health Technology & Innovation SFRN CenterBaltimoreMD
| |
Collapse
|
11
|
Zaleski AL, Berkowsky R, Craig KJT, Pescatello LS. Comprehensiveness, Accuracy, and Readability of Exercise Recommendations Provided by an AI-Based Chatbot: Mixed Methods Study. JMIR MEDICAL EDUCATION 2024; 10:e51308. [PMID: 38206661 PMCID: PMC10811574 DOI: 10.2196/51308] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 10/05/2023] [Accepted: 12/11/2023] [Indexed: 01/12/2024]
Abstract
BACKGROUND Regular physical activity is critical for health and disease prevention. Yet, health care providers and patients face barriers to implement evidence-based lifestyle recommendations. The potential to augment care with the increased availability of artificial intelligence (AI) technologies is limitless; however, the suitability of AI-generated exercise recommendations has yet to be explored. OBJECTIVE The purpose of this study was to assess the comprehensiveness, accuracy, and readability of individualized exercise recommendations generated by a novel AI chatbot. METHODS A coding scheme was developed to score AI-generated exercise recommendations across ten categories informed by gold-standard exercise recommendations, including (1) health condition-specific benefits of exercise, (2) exercise preparticipation health screening, (3) frequency, (4) intensity, (5) time, (6) type, (7) volume, (8) progression, (9) special considerations, and (10) references to the primary literature. The AI chatbot was prompted to provide individualized exercise recommendations for 26 clinical populations using an open-source application programming interface. Two independent reviewers coded AI-generated content for each category and calculated comprehensiveness (%) and factual accuracy (%) on a scale of 0%-100%. Readability was assessed using the Flesch-Kincaid formula. Qualitative analysis identified and categorized themes from AI-generated output. RESULTS AI-generated exercise recommendations were 41.2% (107/260) comprehensive and 90.7% (146/161) accurate, with the majority (8/15, 53%) of inaccuracy related to the need for exercise preparticipation medical clearance. Average readability level of AI-generated exercise recommendations was at the college level (mean 13.7, SD 1.7), with an average Flesch reading ease score of 31.1 (SD 7.7). Several recurring themes and observations of AI-generated output included concern for liability and safety, preference for aerobic exercise, and potential bias and direct discrimination against certain age-based populations and individuals with disabilities. CONCLUSIONS There were notable gaps in the comprehensiveness, accuracy, and readability of AI-generated exercise recommendations. Exercise and health care professionals should be aware of these limitations when using and endorsing AI-based technologies as a tool to support lifestyle change involving exercise.
Collapse
Affiliation(s)
- Amanda L Zaleski
- Clinical Evidence Development, Aetna Medical Affairs, CVS Health Corporation, Hartford, CT, United States
- Department of Preventive Cardiology, Hartford Hospital, Hartford, CT, United States
| | - Rachel Berkowsky
- Department of Kinesiology, University of Connecticut, Storrs, CT, United States
| | - Kelly Jean Thomas Craig
- Clinical Evidence Development, Aetna Medical Affairs, CVS Health Corporation, Hartford, CT, United States
| | - Linda S Pescatello
- Department of Kinesiology, University of Connecticut, Storrs, CT, United States
| |
Collapse
|
12
|
Fazakarley CA, Breen M, Leeson P, Thompson B, Williamson V. Experiences of using artificial intelligence in healthcare: a qualitative study of UK clinician and key stakeholder perspectives. BMJ Open 2023; 13:e076950. [PMID: 38081671 PMCID: PMC10729128 DOI: 10.1136/bmjopen-2023-076950] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/21/2023] [Accepted: 11/24/2023] [Indexed: 12/18/2023] Open
Abstract
OBJECTIVES Artificial intelligence (AI) is a rapidly developing field in healthcare, with tools being developed across various specialties to support healthcare professionals and reduce workloads. It is important to understand the experiences of professionals working in healthcare to ensure that future AI tools are acceptable and effectively implemented. The aim of this study was to gain an in-depth understanding of the experiences and perceptions of UK healthcare workers and other key stakeholders about the use of AI in the National Health Service (NHS). DESIGN A qualitative study using semistructured interviews conducted remotely via MS Teams. Thematic analysis was carried out. SETTING NHS and UK higher education institutes. PARTICIPANTS Thirteen participants were recruited, including clinical and non-clinical participants working for the NHS and researchers working to develop AI tools for healthcare settings. RESULTS Four core themes were identified: positive perceptions of AI; potential barriers to using AI in healthcare; concerns regarding AI use and steps needed to ensure the acceptability of future AI tools. Overall, we found that those working in healthcare were generally open to the use of AI and expected it to have many benefits for patients and facilitate access to care. However, concerns were raised regarding the security of patient data, the potential for misdiagnosis and that AI could increase the burden on already strained healthcare staff. CONCLUSION This study found that healthcare staff are willing to engage with AI research and incorporate AI tools into care pathways. Going forward, the NHS and AI developers will need to collaborate closely to ensure that future tools are suitable for their intended use and do not negatively impact workloads or patient trust. Future AI studies should continue to incorporate the views of key stakeholders to improve tool acceptability. TRIAL REGISTRATION NUMBER NCT05028179; ISRCTN15113915; IRAS ref: 293515.
Collapse
Affiliation(s)
| | - Maria Breen
- School of Psychology & Clinical Language Sciences, University of Reading, Reading, UK
- Breen Clinical Research, London, UK
| | - Paul Leeson
- Division of Cardiovascular Medicine, University of Oxford, Oxford, UK
| | | | - Victoria Williamson
- King's College London, London, UK
- Experimental Psychology, University of Oxford, Oxford, UK
| |
Collapse
|