1
|
Ellison IE, Oslock WM, Abdullah A, Wood L, Thirumalai M, English N, Jones BA, Hollis R, Rubyan M, Chu DI. De novo generation of colorectal patient educational materials using large language models: Prompt engineering key to improved readability. Surgery 2025; 180:109024. [PMID: 39756334 PMCID: PMC11936715 DOI: 10.1016/j.surg.2024.109024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Revised: 10/17/2024] [Accepted: 11/29/2024] [Indexed: 01/07/2025]
Abstract
BACKGROUND Improving patient education has been shown to improve clinical outcomes and reduce disparities, though such efforts can be labor intensive. Large language models may serve as an accessible method to improve patient educational material. The aim of this study was to compare readability between existing educational materials and those generated by large language models. METHODS Baseline colorectal surgery educational materials were gathered from a large academic institution (n = 52). Three prompts were entered into Perplexity and ChatGPT 3.5 for each topic: a Basic prompt that simply requested patient educational information the topic, an Iterative prompt that repeated instruction asking for the information to be more health literate, and a Metric-based prompt that requested a sixth-grade reading level, short sentences, and short words. Flesch-Kincaid Grade Level or Grade Level, Flesch-Kincaid Reading Ease or Ease, and Modified Grade Level scores were calculated for all materials, and unpaired t tests were used to compare mean scores between baseline and documents generated by artificial intelligence platforms. RESULTS Overall existing materials were longer than materials generated by the large language models across categories and prompts: 863-956 words vs 170-265 (ChatGPT) and 220-313 (Perplexity), all P < .01. Baseline materials did not meet sixth-grade readability guidelines based on grade level (Grade Level 7.0-9.8 and Modified Grade Level 9.6-11.5) or ease of readability (Ease 53.1-65.0). Readability of materials generated by a large language model varied by prompt and platform. Overall, ChatGPT materials were more readable than baseline materials with the Metric-based prompt: Grade Level 5.2 vs 8.1, Modified Grade Level 7.3 vs 10.3, and Ease 70.5 vs 60.4, all P < .01. In contrast, Perplexity-generated materials were significantly less readable except for those generated with the Metric-based prompt, which did not statistically differ. CONCLUSION Both existing materials and the majority of educational materials created by large language models did not meet readability recommendations. The exception to this was with ChatGPT materials generated with a Metric-based prompt that consistently improved readability scores from baseline and met recommendations in terms of the average Grade Level score. The variability in performance highlights the importance of the prompt used with large language models.
Collapse
Affiliation(s)
- India E Ellison
- Department of Surgery, University of Alabama at Birmingham, AL
| | - Wendelyn M Oslock
- Department of Surgery, University of Alabama at Birmingham, AL; Department of Quality, Birmingham Veterans Affairs Medical Center, AL. https://www.twitter.com/WendelynOslock
| | - Abiha Abdullah
- Trauma and Transfusion Department, University of Pittsburgh Medical College, PA. https://www.twitter.com/abihaabdullah7
| | - Lauren Wood
- Department of Surgery, University of Alabama at Birmingham, AL
| | | | - Nathan English
- Department of Surgery, University of Alabama at Birmingham, AL; Department of General Surgery, University of Cape Town, WC, South Africa
| | - Bayley A Jones
- Department of Surgery, University of Alabama at Birmingham, AL; Department of Surgery, University of Texas Southwestern Medical Center, Dallas, TX. https://www.twitter.com/bayley_jones
| | - Robert Hollis
- Department of Surgery, University of Alabama at Birmingham, AL. https://www.twitter.com/rhhollis
| | - Michael Rubyan
- University of Michigan School of Public Health, Ann Arbor, MI
| | - Daniel I Chu
- Department of Surgery, University of Alabama at Birmingham, AL.
| |
Collapse
|
2
|
Teixeira da Silva JA, Daly T, Türp JC, Sabel BA, Kendall G. The undeclared use of third-party service providers in academic publishing is unethical: an epistemic reflection and scoping review. NAUNYN-SCHMIEDEBERG'S ARCHIVES OF PHARMACOLOGY 2024; 397:9435-9447. [PMID: 38990307 PMCID: PMC11582143 DOI: 10.1007/s00210-024-03177-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/27/2024] [Accepted: 05/21/2024] [Indexed: 07/12/2024]
Abstract
There is a substantial body of scientific literature on the use of third-party services (TPS) by academics to assist as "publication consultants" in scholarly publishing. TPS provide a wide range of scholarly services to research teams that lack the equipment, skills, motivation, or time to produce a paper without external assistance. While services such as language editing, statistical support, or graphic design are common and often legitimate, some TPS also provide illegitimate services and send unsolicited e-mails (spam) to academics offering these services. Such illegitimate types of TPS have the potential to threaten the integrity of the peer-reviewed scientific literature. In extreme cases, for-profit agencies known as "paper mills" even offer fake scientific publications or authorship slots for sale. The use of such illegitimate services as well as the failure to acknowledge their use is an ethical violation in academic publishing, while the failure to declare support for a TPS can be considered a form of contract fraud. We discuss some literature on TPS, highlight services currently offered by ten of the largest commercial publishers and expect authors to be transparent about the use of these services in their publications. From an ethical/moral (i.e., non-commercial) point of view, it is the responsibility of editors, journals, and publishers, and it should be in their best interest to ensure that illegitimate TPS are identified and prohibited, while publisher-employed TPS should be properly disclosed in their publications.
Collapse
Affiliation(s)
| | - Timothy Daly
- Bioethics Program, FLACSO Argentina, Buenos Aires, Argentina.
- Science Norms Democracy, UMR 8011, Sorbonne Université, Paris, France.
| | - Jens C Türp
- Department of Oral Health & Medicine, University Center for Dental Medicine UZB, University of Basel, Basel, Switzerland.
| | - Bernhard A Sabel
- Institute of Medical Psychology, Medical Faculty, Otto-von-Guericke University of Magdeburg, Leipziger Straße 44, Magdeburg, 39120, Germany.
| | - Graham Kendall
- School of Engineering and Computing, MILA University, No. 1, Persiaran MIU, 71800 Putra Nilai, Negeri Sembilan Darul Khusus, Malaysia.
- School of Computer Science, The University of Nottingham, University Park, Nottingham, NG7 2RD, United Kingdom.
| |
Collapse
|
3
|
Graf EM, McKinney JA, Dye AB, Lin L, Sanchez-Ramos L. Exploring the Limits of Artificial Intelligence for Referencing Scientific Articles. Am J Perinatol 2024; 41:2072-2081. [PMID: 38653452 DOI: 10.1055/s-0044-1786033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 04/25/2024]
Abstract
OBJECTIVE To evaluate the reliability of three artificial intelligence (AI) chatbots (ChatGPT, Google Bard, and Chatsonic) in generating accurate references from existing obstetric literature. STUDY DESIGN Between mid-March and late April 2023, ChatGPT, Google Bard, and Chatsonic were prompted to provide references for specific obstetrical randomized controlled trials (RCTs) published in 2020. RCTs were considered for inclusion if they were mentioned in a previous article that primarily evaluated RCTs published by the top medical and obstetrics and gynecology journals with the highest impact factors in 2020 as well as RCTs published in a new journal focused on publishing obstetric RCTs. The selection of the three AI models was based on their popularity, performance in natural language processing, and public availability. Data collection involved prompting the AI chatbots to provide references according to a standardized protocol. The primary evaluation metric was the accuracy of each AI model in correctly citing references, including authors, publication title, journal name, and digital object identifier (DOI). Statistical analysis was performed using a permutation test to compare the performance of the AI models. RESULTS Among the 44 RCTs analyzed, Google Bard demonstrated the highest accuracy, correctly citing 13.6% of the requested RCTs, whereas ChatGPT and Chatsonic exhibited lower accuracy rates of 2.4 and 0%, respectively. Google Bard often substantially outperformed Chatsonic and ChatGPT in correctly citing the studied reference components. The majority of references from all AI models studied were noted to provide DOIs for unrelated studies or DOIs that do not exist. CONCLUSION To ensure the reliability of scientific information being disseminated, authors must exercise caution when utilizing AI for scientific writing and literature search. However, despite their limitations, collaborative partnerships between AI systems and researchers have the potential to drive synergistic advancements, leading to improved patient care and outcomes. KEY POINTS · AI chatbots often cite scientific articles incorrectly.. · AI chatbots can create false references.. · Responsible AI use in research is vital..
Collapse
Affiliation(s)
- Emily M Graf
- Department of Obstetrics and Gynecology, University of Florida College of Medicine, Jacksonville, Florida
| | - Jordan A McKinney
- Department of Obstetrics and Gynecology, University of Florida College of Medicine, Jacksonville, Florida
| | - Alexander B Dye
- Department of Obstetrics and Gynecology, University of Florida College of Medicine, Jacksonville, Florida
| | - Lifeng Lin
- Department of Epidemiology and Biostatistics, University of Arizona, Tucson, Arizona
| | - Luis Sanchez-Ramos
- Department of Obstetrics and Gynecology, University of Florida College of Medicine, Jacksonville, Florida
| |
Collapse
|
4
|
Ahaley SS, Pandey A, Juneja SK, Gupta TS, Vijayakumar S. ChatGPT in medical writing: A game-changer or a gimmick? Perspect Clin Res 2024; 15:165-171. [PMID: 39583920 PMCID: PMC11584153 DOI: 10.4103/picr.picr_167_23] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2023] [Revised: 08/22/2023] [Accepted: 09/06/2023] [Indexed: 11/26/2024] Open
Abstract
OpenAI's ChatGPT (Generative Pre-trained Transformer) is a chatbot that answers questions and performs writing tasks in a conversational tone. Within months of release, multiple sectors are contemplating the varied applications of this chatbot, including medicine, education, and research, all of which are involved in medical communication and scientific publishing. Medical writers and academics use several artificial intelligence (AI) tools and software for research, literature survey, data analyses, referencing, and writing. There are benefits of using different AI tools in medical writing. However, using chatbots for medical communications pose some major concerns such as potential inaccuracies, data bias, security, and ethical issues. Perceived incorrect notions also limit their use. Moreover, ChatGPT can also be challenging if used incorrectly and for irrelevant tasks. If used appropriately, ChatGPT will not only upgrade the knowledge of the medical writer but also save time and energy that could be directed toward more creative and analytical areas requiring expert skill sets. This review introduces chatbots, outlines the progress in ChatGPT research, elaborates the potential uses of ChatGPT in medical communications along with its challenges and limitations, and proposes future research perspectives. It aims to provide guidance for doctors, researchers, and medical writers on the uses of ChatGPT in medical communications.
Collapse
Affiliation(s)
- Shital Sarah Ahaley
- Hashtag Medical Writing Solutions Private Limited, Chennai, Tamil Nadu, India
| | - Ankita Pandey
- Hashtag Medical Writing Solutions Private Limited, Chennai, Tamil Nadu, India
| | - Simran Kaur Juneja
- Hashtag Medical Writing Solutions Private Limited, Chennai, Tamil Nadu, India
| | - Tanvi Suhane Gupta
- Hashtag Medical Writing Solutions Private Limited, Chennai, Tamil Nadu, India
| | - Sujatha Vijayakumar
- Hashtag Medical Writing Solutions Private Limited, Chennai, Tamil Nadu, India
| |
Collapse
|
5
|
Alyasiri OM, Salman AM, Akhtom D, Salisu S. ChatGPT revisited: Using ChatGPT-4 for finding references and editing language in medical scientific articles. JOURNAL OF STOMATOLOGY, ORAL AND MAXILLOFACIAL SURGERY 2024; 125:101842. [PMID: 38521243 DOI: 10.1016/j.jormas.2024.101842] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Revised: 03/06/2024] [Accepted: 03/20/2024] [Indexed: 03/25/2024]
Abstract
The attainment of academic superiority relies heavily upon the accessibility of scholarly resources and the expression of research findings through faultless language usage. Although modern tools, such as the Publish or Perish software program, are proficient in sourcing academic papers based on specific keywords, they often fall short of extracting comprehensive content, including crucial references. The challenge of linguistic precision remains a prominent issue, particularly for research papers composed by non-native English speakers who may encounter word usage errors. This manuscript serves a twofold purpose: firstly, it reassesses the effectiveness of ChatGPT-4 in the context of retrieving pertinent references tailored to specific research topics. Secondly, it introduces a suite of language editing services that are skilled in rectifying word usage errors, ensuring the refined presentation of research outcomes. The article also provides practical guidelines for formulating precise queries to mitigate the risks of erroneous language usage and the inclusion of spurious references. In the ever-evolving realm of academic discourse, leveraging the potential of advanced AI, such as ChatGPT-4, can significantly enhance the quality and impact of scientific publications.
Collapse
Affiliation(s)
- Osamah Mohammed Alyasiri
- Karbala Technical Institute, Al-Furat Al-Awsat Technical University, Karbala 56001, Iraq; School of Computer Sciences, Universiti Sains Malaysia, Penang 11800, Malaysia.
| | - Amer M Salman
- School of Mathematical Sciences, Universiti Sains Malaysia, Penang 11800, Malaysia
| | - Dua'a Akhtom
- School of Computer Sciences, Universiti Sains Malaysia, Penang 11800, Malaysia
| | - Sani Salisu
- School of Computer Sciences, Universiti Sains Malaysia, Penang 11800, Malaysia; Department of Information Technology, Federal University Dutse, Dutse 720101, Nigeria
| |
Collapse
|
6
|
Bortolotti L, Schauman S, Caligiuri ME. Boosting reproducible research practices with the repeat it with me: Reproducibility team challenge. Magn Reson Med 2024; 92:886-889. [PMID: 38441403 DOI: 10.1002/mrm.30041] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2023] [Revised: 01/18/2024] [Accepted: 01/19/2024] [Indexed: 06/27/2024]
Affiliation(s)
- Laura Bortolotti
- Sir Peter Mansfield Imaging Centre (SPMIC), University of Nottingham, Nottingham, UK
| | - Sophie Schauman
- Department of Clinical Neuro Science, Karolinska Institutet, Solna, Sweden
| | - Maria Eugenia Caligiuri
- Neuroscience Research Center, Department of Medical and Surgical Sciences, Neuroscience Research Center, Università Magna Graecia, Catanzaro, Italy
| |
Collapse
|
7
|
Matsui K, Utsumi T, Aoki Y, Maruki T, Takeshima M, Takaesu Y. Human-Comparable Sensitivity of Large Language Models in Identifying Eligible Studies Through Title and Abstract Screening: 3-Layer Strategy Using GPT-3.5 and GPT-4 for Systematic Reviews. J Med Internet Res 2024; 26:e52758. [PMID: 39151163 PMCID: PMC11364944 DOI: 10.2196/52758] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2023] [Revised: 03/10/2024] [Accepted: 06/25/2024] [Indexed: 08/18/2024] Open
Abstract
BACKGROUND The screening process for systematic reviews is resource-intensive. Although previous machine learning solutions have reported reductions in workload, they risked excluding relevant papers. OBJECTIVE We evaluated the performance of a 3-layer screening method using GPT-3.5 and GPT-4 to streamline the title and abstract-screening process for systematic reviews. Our goal is to develop a screening method that maximizes sensitivity for identifying relevant records. METHODS We conducted screenings on 2 of our previous systematic reviews related to the treatment of bipolar disorder, with 1381 records from the first review and 3146 from the second. Screenings were conducted using GPT-3.5 (gpt-3.5-turbo-0125) and GPT-4 (gpt-4-0125-preview) across three layers: (1) research design, (2) target patients, and (3) interventions and controls. The 3-layer screening was conducted using prompts tailored to each study. During this process, information extraction according to each study's inclusion criteria and optimization for screening were carried out using a GPT-4-based flow without manual adjustments. Records were evaluated at each layer, and those meeting the inclusion criteria at all layers were subsequently judged as included. RESULTS On each layer, both GPT-3.5 and GPT-4 were able to process about 110 records per minute, and the total time required for screening the first and second studies was approximately 1 hour and 2 hours, respectively. In the first study, the sensitivities/specificities of the GPT-3.5 and GPT-4 were 0.900/0.709 and 0.806/0.996, respectively. Both screenings by GPT-3.5 and GPT-4 judged all 6 records used for the meta-analysis as included. In the second study, the sensitivities/specificities of the GPT-3.5 and GPT-4 were 0.958/0.116 and 0.875/0.855, respectively. The sensitivities for the relevant records align with those of human evaluators: 0.867-1.000 for the first study and 0.776-0.979 for the second study. Both screenings by GPT-3.5 and GPT-4 judged all 9 records used for the meta-analysis as included. After accounting for justifiably excluded records by GPT-4, the sensitivities/specificities of the GPT-4 screening were 0.962/0.996 in the first study and 0.943/0.855 in the second study. Further investigation indicated that the cases incorrectly excluded by GPT-3.5 were due to a lack of domain knowledge, while the cases incorrectly excluded by GPT-4 were due to misinterpretations of the inclusion criteria. CONCLUSIONS Our 3-layer screening method with GPT-4 demonstrated acceptable level of sensitivity and specificity that supports its practical application in systematic review screenings. Future research should aim to generalize this approach and explore its effectiveness in diverse settings, both medical and nonmedical, to fully establish its use and operational feasibility.
Collapse
Affiliation(s)
- Kentaro Matsui
- Department of Clinical Laboratory, National Center Hospital, National Center of Neurology and Psychiatry, Kodaira, Japan
- Department of Sleep-Wake Disorders, National Institute of Mental Health, National Center of Neurology and Psychiatry, Kodaira, Japan
| | - Tomohiro Utsumi
- Department of Sleep-Wake Disorders, National Institute of Mental Health, National Center of Neurology and Psychiatry, Kodaira, Japan
- Department of Psychiatry, The Jikei University School of Medicine, Tokyo, Japan
| | - Yumi Aoki
- Graduate School of Nursing Science, St. Luke's International University, Tokyo, Japan
| | - Taku Maruki
- Department of Neuropsychiatry, Kyorin University School of Medicine, Tokyo, Japan
| | - Masahiro Takeshima
- Department of Neuropsychiatry, Akita University Graduate School of Medicine, Akita, Japan
| | - Yoshikazu Takaesu
- Department of Neuropsychiatry, Graduate School of Medicine, University of the Ryukyus, Okinawa, Japan
| |
Collapse
|
8
|
Fatima A, Shafique MA, Alam K, Fadlalla Ahmed TK, Mustafa MS. ChatGPT in medicine: A cross-disciplinary systematic review of ChatGPT's (artificial intelligence) role in research, clinical practice, education, and patient interaction. Medicine (Baltimore) 2024; 103:e39250. [PMID: 39121303 PMCID: PMC11315549 DOI: 10.1097/md.0000000000039250] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 07/19/2024] [Indexed: 08/11/2024] Open
Abstract
BACKGROUND ChatGPT, a powerful AI language model, has gained increasing prominence in medicine, offering potential applications in healthcare, clinical decision support, patient communication, and medical research. This systematic review aims to comprehensively assess the applications of ChatGPT in healthcare education, research, writing, patient communication, and practice while also delineating potential limitations and areas for improvement. METHOD Our comprehensive database search retrieved relevant papers from PubMed, Medline and Scopus. After the screening process, 83 studies met the inclusion criteria. This review includes original studies comprising case reports, analytical studies, and editorials with original findings. RESULT ChatGPT is useful for scientific research and academic writing, and assists with grammar, clarity, and coherence. This helps non-English speakers and improves accessibility by breaking down linguistic barriers. However, its limitations include probable inaccuracy and ethical issues, such as bias and plagiarism. ChatGPT streamlines workflows and offers diagnostic and educational potential in healthcare but exhibits biases and lacks emotional sensitivity. It is useful in inpatient communication, but requires up-to-date data and faces concerns about the accuracy of information and hallucinatory responses. CONCLUSION Given the potential for ChatGPT to transform healthcare education, research, and practice, it is essential to approach its adoption in these areas with caution due to its inherent limitations.
Collapse
Affiliation(s)
- Afia Fatima
- Department of Medicine, Jinnah Sindh Medical University, Karachi, Pakistan
| | | | - Khadija Alam
- Department of Medicine, Liaquat National Medical College, Karachi, Pakistan
| | | | | |
Collapse
|
9
|
Holland AM, Lorenz WR, Cavanagh JC, Smart NJ, Ayuso SA, Scarola GT, Kercher KW, Jorgensen LN, Janis JE, Fischer JP, Heniford BT. Comparison of Medical Research Abstracts Written by Surgical Trainees and Senior Surgeons or Generated by Large Language Models. JAMA Netw Open 2024; 7:e2425373. [PMID: 39093561 PMCID: PMC11297395 DOI: 10.1001/jamanetworkopen.2024.25373] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Accepted: 06/04/2024] [Indexed: 08/04/2024] Open
Abstract
Importance Artificial intelligence (AI) has permeated academia, especially OpenAI Chat Generative Pretrained Transformer (ChatGPT), a large language model. However, little has been reported on its use in medical research. Objective To assess a chatbot's capability to generate and grade medical research abstracts. Design, Setting, and Participants In this cross-sectional study, ChatGPT versions 3.5 and 4.0 (referred to as chatbot 1 and chatbot 2) were coached to generate 10 abstracts by providing background literature, prompts, analyzed data for each topic, and 10 previously presented, unassociated abstracts to serve as models. The study was conducted between August 2023 and February 2024 (including data analysis). Exposure Abstract versions utilizing the same topic and data were written by a surgical trainee or a senior physician or generated by chatbot 1 and chatbot 2 for comparison. The 10 training abstracts were written by 8 surgical residents or fellows, edited by the same senior surgeon, at a high-volume hospital in the Southeastern US with an emphasis on outcomes-based research. Abstract comparison was then based on 10 abstracts written by 5 surgical trainees within the first 6 months of their research year, edited by the same senior author. Main Outcomes and Measures The primary outcome measurements were the abstract grades using 10- and 20-point scales and ranks (first to fourth). Abstract versions by chatbot 1, chatbot 2, junior residents, and the senior author were compared and judged by blinded surgeon-reviewers as well as both chatbot models. Five academic attending surgeons from Denmark, the UK, and the US, with extensive experience in surgical organizations, research, and abstract evaluation served as reviewers. Results Surgeon-reviewers were unable to differentiate between abstract versions. Each reviewer ranked an AI-generated version first at least once. Abstracts demonstrated no difference in their median (IQR) 10-point scores (resident, 7.0 [6.0-8.0]; senior author, 7.0 [6.0-8.0]; chatbot 1, 7.0 [6.0-8.0]; chatbot 2, 7.0 [6.0-8.0]; P = .61), 20-point scores (resident, 14.0 [12.0-7.0]; senior author, 15.0 [13.0-17.0]; chatbot 1, 14.0 [12.0-16.0]; chatbot 2, 14.0 [13.0-16.0]; P = .50), or rank (resident, 3.0 [1.0-4.0]; senior author, 2.0 [1.0-4.0]; chatbot 1, 3.0 [2.0-4.0]; chatbot 2, 2.0 [1.0-3.0]; P = .14). The abstract grades given by chatbot 1 were comparable to the surgeon-reviewers' grades. However, chatbot 2 graded more favorably than the surgeon-reviewers and chatbot 1. Median (IQR) chatbot 2-reviewer grades were higher than surgeon-reviewer grades of all 4 abstract versions (resident, 14.0 [12.0-17.0] vs 16.9 [16.0-17.5]; P = .02; senior author, 15.0 [13.0-17.0] vs 17.0 [16.5-18.0]; P = .03; chatbot 1, 14.0 [12.0-16.0] vs 17.8 [17.5-18.5]; P = .002; chatbot 2, 14.0 [13.0-16.0] vs 16.8 [14.5-18.0]; P = .04). When comparing the grades of the 2 chatbots, chatbot 2 gave higher median (IQR) grades for abstracts than chatbot 1 (resident, 14.0 [13.0-15.0] vs 16.9 [16.0-17.5]; P = .003; senior author, 13.5 [13.0-15.5] vs 17.0 [16.5-18.0]; P = .004; chatbot 1, 14.5 [13.0-15.0] vs 17.8 [17.5-18.5]; P = .003; chatbot 2, 14.0 [13.0-15.0] vs 16.8 [14.5-18.0]; P = .01). Conclusions and Relevance In this cross-sectional study, trained chatbots generated convincing medical abstracts, undifferentiable from resident or senior author drafts. Chatbot 1 graded abstracts similarly to surgeon-reviewers, while chatbot 2 was less stringent. These findings may assist surgeon-scientists in successfully implementing AI in medical research.
Collapse
Affiliation(s)
- Alexis M. Holland
- Division of Gastrointestinal and Minimally Invasive Surgery, Department of Surgery, Atrium Health Carolinas Medical Center, Charlotte, North Carolina
| | - William R. Lorenz
- Division of Gastrointestinal and Minimally Invasive Surgery, Department of Surgery, Atrium Health Carolinas Medical Center, Charlotte, North Carolina
| | - Jack C. Cavanagh
- Department of Economics, Massachusetts Institute of Technology, Cambridge
| | - Neil J. Smart
- Division of Colorectal Surgery, Department of Surgery, Royal Devon & Exeter Hospital, Exeter, Devon, United Kingdom
| | - Sullivan A. Ayuso
- Division of Gastrointestinal and Minimally Invasive Surgery, Department of Surgery, Atrium Health Carolinas Medical Center, Charlotte, North Carolina
| | - Gregory T. Scarola
- Division of Gastrointestinal and Minimally Invasive Surgery, Department of Surgery, Atrium Health Carolinas Medical Center, Charlotte, North Carolina
| | - Kent W. Kercher
- Division of Gastrointestinal and Minimally Invasive Surgery, Department of Surgery, Atrium Health Carolinas Medical Center, Charlotte, North Carolina
| | - Lars N. Jorgensen
- Department of Clinical Medicine, University of Copenhagen, Bispedjerg & Frederiksberg Hospital, Copenhagen, Denmark
| | - Jeffrey E. Janis
- Division of Plastic and Reconstructive Surgery, The Ohio State University Wexner Medical Center, Columbus
| | - John P. Fischer
- Division of Plastic Surgery, University of Pennsylvania Health System, Philadelphia
| | - B. Todd Heniford
- Division of Gastrointestinal and Minimally Invasive Surgery, Department of Surgery, Atrium Health Carolinas Medical Center, Charlotte, North Carolina
| |
Collapse
|
10
|
Abi-Rafeh J, Henry N, Xu HH, Bassiri-Tehrani B, Arezki A, Kazan R, Gilardino MS, Nahai F. Utility and Comparative Performance of Current Artificial Intelligence Large Language Models as Postoperative Medical Support Chatbots in Aesthetic Surgery. Aesthet Surg J 2024; 44:889-896. [PMID: 38318684 DOI: 10.1093/asj/sjae025] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2023] [Revised: 01/24/2024] [Accepted: 01/26/2024] [Indexed: 02/07/2024] Open
Abstract
BACKGROUND Large language models (LLMs) have revolutionized the way plastic surgeons and their patients can access and leverage artificial intelligence (AI). OBJECTIVES The present study aims to compare the performance of 2 current publicly available and patient-accessible LLMs in the potential application of AI as postoperative medical support chatbots in an aesthetic surgeon's practice. METHODS Twenty-two simulated postoperative patient presentations following aesthetic breast plastic surgery were devised and expert-validated. Complications varied in their latency within the postoperative period, as well as urgency of required medical attention. In response to each patient-reported presentation, Open AI's ChatGPT and Google's Bard, in their unmodified and freely available versions, were objectively assessed for their comparative accuracy in generating an appropriate differential diagnosis, most-likely diagnosis, suggested medical disposition, treatments or interventions to begin from home, and/or red flag signs/symptoms indicating deterioration. RESULTS ChatGPT cumulatively and significantly outperformed Bard across all objective assessment metrics examined (66% vs 55%, respectively; P < .05). Accuracy in generating an appropriate differential diagnosis was 61% for ChatGPT vs 57% for Bard (P = .45). ChatGPT asked an average of 9.2 questions on history vs Bard's 6.8 questions (P < .001), with accuracies of 91% vs 68% reporting the most-likely diagnosis, respectively (P < .01). Appropriate medical dispositions were suggested with accuracies of 50% by ChatGPT vs 41% by Bard (P = .40); appropriate home interventions/treatments with accuracies of 59% vs 55% (P = .94), and red flag signs/symptoms with accuracies of 79% vs 54% (P < .01), respectively. Detailed and comparative performance breakdowns according to complication latency and urgency are presented. CONCLUSIONS ChatGPT represents the superior LLM for the potential application of AI technology in postoperative medical support chatbots. Imperfect performance and limitations discussed may guide the necessary refinement to facilitate adoption.
Collapse
|
11
|
Luo X, Chen F, Zhu D, Wang L, Wang Z, Liu H, Lyu M, Wang Y, Wang Q, Chen Y. Potential Roles of Large Language Models in the Production of Systematic Reviews and Meta-Analyses. J Med Internet Res 2024; 26:e56780. [PMID: 38819655 PMCID: PMC11234072 DOI: 10.2196/56780] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2024] [Revised: 05/21/2024] [Accepted: 05/29/2024] [Indexed: 06/01/2024] Open
Abstract
Large language models (LLMs) such as ChatGPT have become widely applied in the field of medical research. In the process of conducting systematic reviews, similar tools can be used to expedite various steps, including defining clinical questions, performing the literature search, document screening, information extraction, and language refinement, thereby conserving resources and enhancing efficiency. However, when using LLMs, attention should be paid to transparent reporting, distinguishing between genuine and false content, and avoiding academic misconduct. In this viewpoint, we highlight the potential roles of LLMs in the creation of systematic reviews and meta-analyses, elucidating their advantages, limitations, and future research directions, aiming to provide insights and guidance for authors planning systematic reviews and meta-analyses.
Collapse
Affiliation(s)
- Xufei Luo
- Evidence-Based Medicine Center, School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
- World Health Organization Collaboration Center for Guideline Implementation and Knowledge Translation, Lanzhou, China
- Institute of Health Data Science, Lanzhou University, Lanzhou, China
- Key Laboratory of Evidence Based Medicine and Knowledge Translation of Gansu Province, Lanzhou University, Lanzhou, China
- Research Unit of Evidence-Based Evaluation and Guidelines, Chinese Academy of Medical Sciences (2021RU017), School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
| | - Fengxian Chen
- School of Information Science & Engineering, Lanzhou University, Lanzhou, China
| | - Di Zhu
- School of Public Health, Lanzhou University, Lanzhou, China
| | - Ling Wang
- School of Public Health, Lanzhou University, Lanzhou, China
| | - Zijun Wang
- Evidence-Based Medicine Center, School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
- World Health Organization Collaboration Center for Guideline Implementation and Knowledge Translation, Lanzhou, China
- Institute of Health Data Science, Lanzhou University, Lanzhou, China
- Key Laboratory of Evidence Based Medicine and Knowledge Translation of Gansu Province, Lanzhou University, Lanzhou, China
- Research Unit of Evidence-Based Evaluation and Guidelines, Chinese Academy of Medical Sciences (2021RU017), School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
| | - Hui Liu
- Evidence-Based Medicine Center, School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
- World Health Organization Collaboration Center for Guideline Implementation and Knowledge Translation, Lanzhou, China
- Institute of Health Data Science, Lanzhou University, Lanzhou, China
- Key Laboratory of Evidence Based Medicine and Knowledge Translation of Gansu Province, Lanzhou University, Lanzhou, China
- Research Unit of Evidence-Based Evaluation and Guidelines, Chinese Academy of Medical Sciences (2021RU017), School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
| | - Meng Lyu
- School of Public Health, Lanzhou University, Lanzhou, China
| | - Ye Wang
- School of Public Health, Lanzhou University, Lanzhou, China
| | - Qi Wang
- Department of Health Research Methods, Evidence and Impact, Faculty of Health Sciences, McMaster University, Hamilton, ON, Canada
- McMaster Health Forum, McMaster University, Hamilton, ON, Canada
| | - Yaolong Chen
- Evidence-Based Medicine Center, School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
- World Health Organization Collaboration Center for Guideline Implementation and Knowledge Translation, Lanzhou, China
- Institute of Health Data Science, Lanzhou University, Lanzhou, China
- Key Laboratory of Evidence Based Medicine and Knowledge Translation of Gansu Province, Lanzhou University, Lanzhou, China
- Research Unit of Evidence-Based Evaluation and Guidelines, Chinese Academy of Medical Sciences (2021RU017), School of Basic Medical Sciences, Lanzhou University, Lanzhou, China
| |
Collapse
|
12
|
González R, Poenaru D, Woo R, Trappey AF, Carter S, Darcy D, Encisco E, Gulack B, Miniati D, Tombash E, Huang EY. ChatGPT: What Every Pediatric Surgeon Should Know About Its Potential Uses and Pitfalls. J Pediatr Surg 2024; 59:941-947. [PMID: 38336588 DOI: 10.1016/j.jpedsurg.2024.01.007] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/02/2023] [Revised: 12/30/2023] [Accepted: 01/09/2024] [Indexed: 02/12/2024]
Abstract
ChatGPT - currently the most popular generative artificial intelligence system - has been revolutionizing the world and healthcare since its release in November 2022. ChatGPT is a conversational chatbot that uses machine learning algorithms to enhance its replies based on user interactions and is a part of a broader effort to develop natural language processing that can assist people in their daily lives by understanding and responding to human language in a useful and engaging way. Thus far, many potential applications within healthcare have been described, despite its relatively recent release. This manuscript offers the pediatric surgical community a primer on this new technology and discusses some initial observations about its potential uses and pitfalls. Moreover, it introduces the perspectives of medical journals and surgical societies regarding the use of this artificial intelligence chatbot. As ChatGPT and other large language models continue to evolve, it is the responsibility of the pediatric surgery community to stay abreast of these changes and play an active role in safely incorporating them into our field for the benefit of our patients. LEVEL OF EVIDENCE: V.
Collapse
Affiliation(s)
- Raquel González
- Division of Pediatric Surgery, Johns Hopkins All Children's Hospital, 501 6th Avenue S, Saint Petersburg, FL, 33701, USA.
| | - Dan Poenaru
- McGill University, 5252 Boul. De Maissonneuve O. rm. 3E.05, Montréal, QC, H4a 3S5, Canada
| | - Russell Woo
- Department of Surgery, Division of Pediatric Surgery, University of Hawai'i, John A. Burns School of Medicine, 1319 Punahou Street, Suite 600, Honolulu, HI, 96826, USA
| | - A Francois Trappey
- Pediatric General and Thoracic Surgery, Brooke Army Medical Center, 3551 Roger Brooke Dr, Fort Sam Houston, TX, 78234, USA
| | - Stewart Carter
- Division of Pediatric Surgery, University of Louisville, Norton Children's Hospital, 315 East Broadway, Suite 565, Louisville, KY, 40202, USA
| | - David Darcy
- Golisano Children's Hospital, University of Rochester Medical Center, 601 Elmwood Avenue, Box SURG, Rochester, NY, 14642, USA
| | - Ellen Encisco
- Division of Pediatric General and Thoracic Surgery, Cincinnati Children's Hospital, 3333 Burnet Ave, Cincinnati, OH, 45229, USA
| | - Brian Gulack
- Rush University Medical Center, 1653 W Congress Parkway, Kellogg, Chicago, IL, 60612, USA
| | - Doug Miniati
- Department of Pediatric Surgery, Kaiser Permanente Roseville, 1600 Eureka Road, Building C, Suite C35, Roseville, CA, 95661, USA
| | - Edzhem Tombash
- Division of Pediatric General and Thoracic Surgery, Cincinnati Children's Hospital, 3333 Burnet Ave, Cincinnati, OH, 45229, USA
| | - Eunice Y Huang
- Vanderbilt University Medical Center, Monroe Carell Jr. Children's Hospital, 2200 Children's Way, Suite 7100, Nashville, TN, 37232, USA
| |
Collapse
|
13
|
Alapati R, Campbell D, Molin N, Creighton E, Wei Z, Boon M, Huntley C. Evaluating insomnia queries from an artificial intelligence chatbot for patient education. J Clin Sleep Med 2024; 20:583-594. [PMID: 38217478 PMCID: PMC10985291 DOI: 10.5664/jcsm.10948] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2023] [Revised: 11/28/2023] [Accepted: 11/28/2023] [Indexed: 01/15/2024]
Abstract
STUDY OBJECTIVES We evaluated the accuracy of ChatGPT in addressing insomnia-related queries for patient education and assessed ChatGPT's ability to provide varied responses based on differing prompting scenarios. METHODS Four identical sets of 20 insomnia-related queries were posed to ChatGPT. Each set differed by the context in which ChatGPT was prompted: no prompt, patient-centered, physician-centered, and with references and statistics. Responses were reviewed by 2 academic sleep surgeons, 1 academic sleep medicine physician, and 2 sleep medicine fellows across 4 domains: clinical accuracy, prompt adherence, referencing, and statistical precision, using a binary grading system. Flesch-Kincaid grade-level scores were calculated to estimate the grade level of the responses, with statistical differences between prompts analyzed via analysis of variance and Tukey's test. Interrater reliability was calculated using Fleiss's kappa. RESULTS The study revealed significant variations in the Flesch-Kincaid grade-level scores across 4 prompts: unprompted (13.2 ± 2.2), patient-centered (8.1 ± 1.9), physician-centered (15.4 ± 2.8), and with references and statistics (17.3 ± 2.3, P < .001). Despite poor Fleiss kappa scores, indicating low interrater reliability for clinical accuracy and relevance, all evaluators agreed that the majority of ChatGPT's responses were clinically accurate, with the highest variability on Form 4. The responses were also uniformly relevant to the given prompts (100% agreement). Eighty percent of the references ChatGPT cited were verified as both real and relevant, and only 25% of cited statistics were corroborated within referenced articles. CONCLUSIONS ChatGPT can be used to generate clinically accurate responses to insomnia-related inquiries. CITATION Alapati R, Campbell D, Molin N, et al. Evaluating insomnia queries from an artificial intelligence chatbot for patient education. J Clin Sleep Med. 2024;20(4):583-594.
Collapse
Affiliation(s)
- Rahul Alapati
- Sidney Kimmel Medical College, Thomas Jefferson University, Philadelphia, Pennsylvania
| | - Daniel Campbell
- Department of Otolaryngology, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
| | - Nicole Molin
- Department of Otolaryngology, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
- Department of Neurology, Jefferson Sleep Disorders Center, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
| | - Erin Creighton
- Department of Otolaryngology, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
- Department of Neurology, Jefferson Sleep Disorders Center, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
| | - Zhikui Wei
- Department of Neurology, Jefferson Sleep Disorders Center, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
| | - Maurits Boon
- Department of Otolaryngology, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
| | - Colin Huntley
- Department of Otolaryngology, Thomas Jefferson University Hospitals, Philadelphia, Pennsylvania
| |
Collapse
|
14
|
Abi-Rafeh J, Xu HH, Kazan R, Tevlin R, Furnas H. Large Language Models and Artificial Intelligence: A Primer for Plastic Surgeons on the Demonstrated and Potential Applications, Promises, and Limitations of ChatGPT. Aesthet Surg J 2024; 44:329-343. [PMID: 37562022 DOI: 10.1093/asj/sjad260] [Citation(s) in RCA: 15] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2023] [Revised: 08/02/2023] [Accepted: 08/04/2023] [Indexed: 08/12/2023] Open
Abstract
BACKGROUND The rapidly evolving field of artificial intelligence (AI) holds great potential for plastic surgeons. ChatGPT, a recently released AI large language model (LLM), promises applications across many disciplines, including healthcare. OBJECTIVES The aim of this article was to provide a primer for plastic surgeons on AI, LLM, and ChatGPT, including an analysis of current demonstrated and proposed clinical applications. METHODS A systematic review was performed identifying medical and surgical literature on ChatGPT's proposed clinical applications. Variables assessed included applications investigated, command tasks provided, user input information, AI-emulated human skills, output validation, and reported limitations. RESULTS The analysis included 175 articles reporting on 13 plastic surgery applications and 116 additional clinical applications, categorized by field and purpose. Thirty-four applications within plastic surgery are thus proposed, with relevance to different target audiences, including attending plastic surgeons (n = 17, 50%), trainees/educators (n = 8, 24.0%), researchers/scholars (n = 7, 21%), and patients (n = 2, 6%). The 15 identified limitations of ChatGPT were categorized by training data, algorithm, and ethical considerations. CONCLUSIONS Widespread use of ChatGPT in plastic surgery will depend on rigorous research of proposed applications to validate performance and address limitations. This systemic review aims to guide research, development, and regulation to safely adopt AI in plastic surgery.
Collapse
|
15
|
Vieira AGDS, Saconato H, Eid RAC, Nawa RK. ChatGPT: immutable insertion in health research and researchers' lives. EINSTEIN-SAO PAULO 2024; 22:eCE0752. [PMID: 38477797 PMCID: PMC11730319 DOI: 10.31744/einstein_journal/2024ce0752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Accepted: 10/18/2023] [Indexed: 03/14/2024] Open
|
16
|
Zybaczynska J, Norris M, Modi S, Brennan J, Jhaveri P, Craig TJ, Al-Shaikhly T. Artificial Intelligence-Generated Scientific Literature: A Critical Appraisal. THE JOURNAL OF ALLERGY AND CLINICAL IMMUNOLOGY. IN PRACTICE 2024; 12:106-110. [PMID: 37832818 DOI: 10.1016/j.jaip.2023.10.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/25/2023] [Revised: 08/14/2023] [Accepted: 10/03/2023] [Indexed: 10/15/2023]
Abstract
BACKGROUND Review articles play a critical role in informing medical decisions and identifying avenues for future research. With the introduction of artificial intelligence (AI), there has been a growing interest in the potential of this technology to transform the synthesis of medical literature. Open AI's Generative Pre-trained Transformer (GPT-4) (Open AI Inc, San Francisco, CA) tool provides access to advanced AI that is able to quickly produce medical literature following only simple prompts. The accuracy of the generated articles requires review, especially in subspecialty fields like Allergy/Immunology. OBJECTIVE To critically appraise AI-synthesized allergy-focused minireviews. METHODS We tasked the GPT-4 Chatbot with generating 2 1,000-word reviews on the topics of hereditary angioedema and eosinophilic esophagitis. Authors critically appraised these articles using the Joanna Briggs Institute (JBI) tool for text and opinion and additionally evaluated domains of interest such as language, reference quality, and accuracy of the content. RESULTS The language of the AI-generated minireviews was carefully articulated and logically focused on the topic of interest; however, reviewers of the AI-generated articles indicated that the AI-generated content lacked depth, did not appear to be the result of an analytical process, missed critical information, and contained inaccurate information. Despite being provided instruction to utilize scientific references, the AI chatbot relied mainly on freely available resources, and the AI chatbot fabricated references. CONCLUSIONS The AI holds the potential to change the landscape of synthesizing medical literature; however, apparent inaccurate and fabricated information calls for rigorous evaluation and validation of AI tools in generating medical literature, especially on subjects associated with limited resources.
Collapse
Affiliation(s)
- Justyna Zybaczynska
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Matthew Norris
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Sunjay Modi
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Jennifer Brennan
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Pooja Jhaveri
- Division of Allergy & Immunology, Department of Pediatrics, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Timothy J Craig
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa
| | - Taha Al-Shaikhly
- Section of Allergy, Asthma & Immunology, Department of Medicine, Pennsylvania State University College of Medicine, Hershey, Pa.
| |
Collapse
|
17
|
Sumbal A, Sumbal R, Amir A. Can ChatGPT-3.5 Pass a Medical Exam? A Systematic Review of ChatGPT's Performance in Academic Testing. JOURNAL OF MEDICAL EDUCATION AND CURRICULAR DEVELOPMENT 2024; 11:23821205241238641. [PMID: 38487300 PMCID: PMC10938614 DOI: 10.1177/23821205241238641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 02/25/2024] [Indexed: 03/17/2024]
Abstract
OBJECTIVE We, therefore, aim to conduct a systematic review to assess the academic potential of ChatGPT-3.5, along with its strengths and limitations when giving medical exams. METHOD Following PRISMA guidelines, a systemic search of the literature was performed using electronic databases PUBMED/MEDLINE, Google Scholar, and Cochrane. Articles from their inception till April 4, 2023, were queried. A formal narrative analysis was conducted by systematically arranging similarities and differences between individual findings together. RESULTS After rigorous screening, 12 articles underwent this review. All the selected papers assessed the academic performance of ChatGPT-3.5. One study compared the performance of ChatGPT-3.5 with the performance of ChatGPT-4 when giving a medical exam. Overall, ChatGPT performed well in 4 tests, averaged in 4 tests, and performed badly in 4 tests. ChatGPT's performance was directly proportional to the level of the questions' difficulty but was unremarkable on whether the questions were binary, descriptive, or MCQ-based. ChatGPT's explanation, reasoning, memory, and accuracy were remarkably good, whereas it failed to understand image-based questions, and lacked insight and critical thinking. CONCLUSION ChatGPT-3.5 performed satisfactorily in the exams it took as an examinee. However, there is a need for future related studies to fully explore the potential of ChatGPT in medical education.
Collapse
Affiliation(s)
- Anusha Sumbal
- Dow University of Health Sciences, Karachi, Pakistan
| | - Ramish Sumbal
- Dow University of Health Sciences, Karachi, Pakistan
| | - Alina Amir
- Dow University of Health Sciences, Karachi, Pakistan
| |
Collapse
|
18
|
Kayaalp ME, Ollivier M, Winkler PW, Dahmen J, Musahl V, Hirschmann MT, Karlsson J. Embrace responsible ChatGPT usage to overcome language barriers in academic writing. Knee Surg Sports Traumatol Arthrosc 2024; 32:5-9. [PMID: 38226673 DOI: 10.1002/ksa.12014] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Accepted: 11/08/2023] [Indexed: 01/17/2024]
Affiliation(s)
- M Enes Kayaalp
- Department of Orthopaedic Surgery, University of Pittsburgh, Pittsburgh, Pennsylvania, USA
- Department for Orthopaedics and Traumatology, Istanbul Kartal Research and Training Hospital, Istanbul, Turkiye
| | - Matthieu Ollivier
- CNRS, Institute of Movement Sciences (ISM), Aix Marseille University, Marseille, France
| | - Philipp W Winkler
- Department for Orthopaedics and Traumatology, Kepler University Hospital GmbH, Linz, Austria
| | - Jari Dahmen
- Department of Orthopaedic Surgery and Sports Medicine, Amsterdam Movement Sciences, Amsterdam UMC, University of Amsterdam, Amsterdam, The Netherlands
- Academic Center for Evidence Based Sports Medicine (ACES), Amsterdam, The Netherlands
- Amsterdam Collaboration for Health and Safety in Sports (ACHSS), International Olympic Committee (IOC) Research Center Amsterdam UMC, Amsterdam, The Netherlands
| | - Volker Musahl
- Department of Orthopaedic Surgery, University of Pittsburgh, Pittsburgh, Pennsylvania, USA
| | - Michael T Hirschmann
- Department of Orthopedic Surgery and Traumatology, Head Knee Surgery and DKF Head of Research, Kantonsspital Baselland, Bruderholz, Bottmingen, Switzerland
- University of Basel, Basel, Switzerland
| | - Jon Karlsson
- Department for Orthopaedics, Sahlgrenska University Hospital, Institute of Clinical Sciences, Sahlgrenska Academy, Gothenburg University, Gothenburg, Sweden
| |
Collapse
|
19
|
Puladi B, Gsaxner C, Kleesiek J, Hölzle F, Röhrig R, Egger J. The impact and opportunities of large language models like ChatGPT in oral and maxillofacial surgery: a narrative review. Int J Oral Maxillofac Surg 2024; 53:78-88. [PMID: 37798200 DOI: 10.1016/j.ijom.2023.09.005] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2023] [Revised: 09/14/2023] [Accepted: 09/19/2023] [Indexed: 10/07/2023]
Abstract
Since its release at the end of 2022, the social response to ChatGPT, a large language model (LLM), has been huge, as it has revolutionized the way we communicate with computers. This review was performed to describe the technical background of LLMs and to provide a review of the current literature on LLMs in the field of oral and maxillofacial surgery (OMS). The PubMed, Scopus, and Web of Science databases were searched for LLMs and OMS. Adjacent surgical disciplines were included to cover the entire literature, and records from Google Scholar and medRxiv were added. Out of the 57 records identified, 37 were included; 31 (84%) were related to GPT-3.5, four (11%) to GPT-4, and two (5%) to both. Current research on LLMs is mainly limited to research and scientific writing, patient information/communication, and medical education. Classic OMS diseases are underrepresented. The current literature related to LLMs in OMS has a limited evidence level. There is a need to investigate the use of LLMs scientifically and systematically in the core areas of OMS. Although LLMs are likely to add value outside the operating room, the use of LLMs raises ethical and medical regulatory issues that must first be addressed.
Collapse
Affiliation(s)
- B Puladi
- Department of Oral and Maxillofacial Surgery, University Hospital RWTH Aachen, Aachen, Germany; Institute of Medical Informatics, University Hospital RWTH Aachen, Aachen, Germany
| | - C Gsaxner
- Department of Oral and Maxillofacial Surgery, University Hospital RWTH Aachen, Aachen, Germany; Institute of Medical Informatics, University Hospital RWTH Aachen, Aachen, Germany; Institute of Computer Graphics and Vision, Graz University of Technology, Graz, Austria; Department of Oral and Maxillofacial Surgery, Medical University of Graz, Graz, Austria
| | - J Kleesiek
- Institute for AI in Medicine (IKIM), University Hospital Essen (AöR), Essen, Germany
| | - F Hölzle
- Department of Oral and Maxillofacial Surgery, University Hospital RWTH Aachen, Aachen, Germany
| | - R Röhrig
- Institute of Medical Informatics, University Hospital RWTH Aachen, Aachen, Germany
| | - J Egger
- Institute of Computer Graphics and Vision, Graz University of Technology, Graz, Austria; Institute for AI in Medicine (IKIM), University Hospital Essen (AöR), Essen, Germany.
| |
Collapse
|
20
|
Alkhaaldi SMI, Kassab CH, Dimassi Z, Oyoun Alsoud L, Al Fahim M, Al Hageh C, Ibrahim H. Medical Student Experiences and Perceptions of ChatGPT and Artificial Intelligence: Cross-Sectional Study. JMIR MEDICAL EDUCATION 2023; 9:e51302. [PMID: 38133911 PMCID: PMC10770787 DOI: 10.2196/51302] [Citation(s) in RCA: 25] [Impact Index Per Article: 12.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 11/10/2023] [Accepted: 12/11/2023] [Indexed: 12/23/2023]
Abstract
BACKGROUND Artificial intelligence (AI) has the potential to revolutionize the way medicine is learned, taught, and practiced, and medical education must prepare learners for these inevitable changes. Academic medicine has, however, been slow to embrace recent AI advances. Since its launch in November 2022, ChatGPT has emerged as a fast and user-friendly large language model that can assist health care professionals, medical educators, students, trainees, and patients. While many studies focus on the technology's capabilities, potential, and risks, there is a gap in studying the perspective of end users. OBJECTIVE The aim of this study was to gauge the experiences and perspectives of graduating medical students on ChatGPT and AI in their training and future careers. METHODS A cross-sectional web-based survey of recently graduated medical students was conducted in an international academic medical center between May 5, 2023, and June 13, 2023. Descriptive statistics were used to tabulate variable frequencies. RESULTS Of 325 applicants to the residency programs, 265 completed the survey (an 81.5% response rate). The vast majority of respondents denied using ChatGPT in medical school, with 20.4% (n=54) using it to help complete written assessments and only 9.4% using the technology in their clinical work (n=25). More students planned to use it during residency, primarily for exploring new medical topics and research (n=168, 63.4%) and exam preparation (n=151, 57%). Male students were significantly more likely to believe that AI will improve diagnostic accuracy (n=47, 51.7% vs n=69, 39.7%; P=.001), reduce medical error (n=53, 58.2% vs n=71, 40.8%; P=.002), and improve patient care (n=60, 65.9% vs n=95, 54.6%; P=.007). Previous experience with AI was significantly associated with positive AI perception in terms of improving patient care, decreasing medical errors and misdiagnoses, and increasing the accuracy of diagnoses (P=.001, P<.001, P=.008, respectively). CONCLUSIONS The surveyed medical students had minimal formal and informal experience with AI tools and limited perceptions of the potential uses of AI in health care but had overall positive views of ChatGPT and AI and were optimistic about the future of AI in medical education and health care. Structured curricula and formal policies and guidelines are needed to adequately prepare medical learners for the forthcoming integration of AI in medicine.
Collapse
Affiliation(s)
- Saif M I Alkhaaldi
- Khalifa University College of Medicine and Health Sciences, Abu Dhabi, United Arab Emirates
| | - Carl H Kassab
- Khalifa University College of Medicine and Health Sciences, Abu Dhabi, United Arab Emirates
| | - Zakia Dimassi
- Department of Medical Science, Khalifa University College of Medicine and Health Sciences, Abu Dhabi, United Arab Emirates
| | - Leen Oyoun Alsoud
- Department of Medical Science, Khalifa University College of Medicine and Health Sciences, Abu Dhabi, United Arab Emirates
| | - Maha Al Fahim
- Education Institute, Sheikh Khalifa Medical City, Abu Dhabi, United Arab Emirates
| | - Cynthia Al Hageh
- Department of Medical Science, Khalifa University College of Medicine and Health Sciences, Abu Dhabi, United Arab Emirates
| | - Halah Ibrahim
- Department of Medical Science, Khalifa University College of Medicine and Health Sciences, Abu Dhabi, United Arab Emirates
| |
Collapse
|
21
|
Lingard L, Chandritilake M, de Heer M, Klasen J, Maulina F, Olmos-Vega F, St-Onge C. Will ChatGPT's Free Language Editing Service Level the Playing Field in Science Communication?: Insights from a Collaborative Project with Non-native English Scholars. PERSPECTIVES ON MEDICAL EDUCATION 2023; 12:565-574. [PMID: 38163049 PMCID: PMC10756157 DOI: 10.5334/pme.1246] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Accepted: 11/28/2023] [Indexed: 01/03/2024]
Abstract
ChatGPT has been widely heralded as a way to level the playing field in scientific communication through its free language editing service. However, such claims lack systematic evidence. A writing scholar (LL) and six non-native English scholars researching health professions education collaborated on this Writer's Craft to fill this gap. Our overarching aim was to provide experiential evidence about ChatGPT's performance as a language editor and writing coach. We implemented three cycles of a systematic procedure, describing how we developed our prompts, selected text for editing, incrementally prompted to refine ChatGPT's responses, and analyzed the quality of its language edits and explanations. From this experience, we offer five insights, and we conclude that the optimism about ChatGPT's capacity to level the playing field for non-native English writers should be tempered. In the writer's craft section we offer simple tips to improve your writing in one of three areas: Energy, Clarity and Persuasiveness. Each entry focuses on a key writing feature or strategy, illustrates how it commonly goes wrong, teaches the grammatical underpinnings necessary to understand it and offers suggestions to wield it effectively. We encourage readers to share comments on or suggestions for this section on Twitter, using the hashtag: #how'syourwriting?
Collapse
Affiliation(s)
- Lorelei Lingard
- Centre for Education Research & Innovation, Schulich School of Medicine & Dentistry, Western University, London, Ontario, Canada
| | | | - Merel de Heer
- Amsterdam UMC, Vrije Universiteit Amsterdam, Research in Medical Education, Amsterdam, Netherlands
| | - Jennifer Klasen
- University Digestive Health Care Center, Department of Visceral Surgery, St. Clara Hospital and University Hospital Basel, Switzerland
| | - Fury Maulina
- Faculty of Health, Medicine and Life Sciences, Maastricht University, Maastricht, the Netherlands
- Department of Public Health, Faculty of Medicine, Universitas Malikussaleh, Lhokseumawe, Aceh, Indonesia
| | - Francisco Olmos-Vega
- Department of Medicine and Health Profession Education Center, Universitéde Sherbrooke, Sherbrooke, Canada
| | - Christina St-Onge
- Department of Medicine and Health Profession Education Center, Universitéde Sherbrooke, Sherbrooke, Canada
| |
Collapse
|
22
|
Sahari Y, Al-Kadi AMT, Ali JKM. A Cross Sectional Study of ChatGPT in Translation: Magnitude of Use, Attitudes, and Uncertainties. JOURNAL OF PSYCHOLINGUISTIC RESEARCH 2023; 52:2937-2954. [PMID: 37934302 DOI: 10.1007/s10936-023-10031-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 10/25/2023] [Indexed: 11/08/2023]
Abstract
This preliminary cross-sectional study, focusing on Artificial Intelligence (AI), aimed to assess the impact of ChatGPT on translation within an Arab context. It primarily explored the attitudes of a sample of translation teachers and students through semi-structured interviews and projective techniques. Data collection included gathering information about the advantages and challenges that ChatGPT, in comparison to Google Translate, had introduced to the field of translation and translation teaching. The results indicated that nearly all the participants were satisfied with ChatGPT. The results also revealed that most students preferred ChatGPT over Google Translate, while most teachers favored Google Translate. The study also found that the participants recognized both positive and negative aspects of using ChatGPT in translation. Findings also indicated that ChatGPT, as a recent AI-based translation-related technology, is more valuable for mechanical processes of writing and editing translated texts than for tasks requiring judgment, such as fine-tuning and double-checking. While it offers various advantages, AI also presents new challenges that educators and stakeholders need to address accordingly.
Collapse
|
23
|
Silva TP, Ocampo TSC, Alencar-Palha C, Oliveira-Santos C, Takeshita WM, Oliveira ML. ChatGPT: a tool for scientific writing or a threat to integrity? Br J Radiol 2023; 96:20230430. [PMID: 37750843 PMCID: PMC10646664 DOI: 10.1259/bjr.20230430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2023] [Revised: 08/03/2023] [Accepted: 08/06/2023] [Indexed: 09/27/2023] Open
Abstract
The use of ChatGPT as a tool for writing and knowledge integration raises concerns about the potential for its use to replace critical thinking and academic writing skills. While ChatGPT can assist in generating text and suggesting appropriate language, it should not replace the human responsibility for creating innovative knowledge through experiential learning. The accuracy and quality of information provided by ChatGPT also require caution, as previous studies have reported inaccuracies in references used by chatbots. ChatGPT acknowledges certain limitations, including the potential for generating erroneous or biased content, and it is essential to exercise caution in interpreting its responses and recognize the indispensable role of human experience in the processes of information retrieval and knowledge creation. Furthermore, the challenge of distinguishing between papers written by humans or AI highlights the need for thorough review processes to prevent the spread of articles that could lead to the loss of confidence in the accuracy and integrity of scientific research. Overall, while the use of ChatGPT can be helpful, it is crucial to raise awareness of the potential issues associated with the use of ChatGPT, as well as to discuss boundaries so that AI can be used without compromising the quality of scientific articles and the integrity of evidence-based knowledge.
Collapse
Affiliation(s)
- Thaísa Pinheiro Silva
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| | - Thaís S C Ocampo
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| | - Caio Alencar-Palha
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| | - Christiano Oliveira-Santos
- Department of Diagnosis and Oral Health, University of Louisville School of Dentistry, Louisville, United States
| | - Wilton Mitsunari Takeshita
- Department of Diagnosis and Surgery, Paulista State University Júlio de Mesquita Filho, Araçatuba, Brazil
| | - Matheus L Oliveira
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| |
Collapse
|
24
|
Peacock J, Austin A, Shapiro M, Battista A, Samuel A. Accelerating medical education with ChatGPT: an implementation guide. MEDEDPUBLISH 2023; 13:64. [PMID: 38440148 PMCID: PMC10910173 DOI: 10.12688/mep.19732.2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/17/2023] [Indexed: 03/06/2024] Open
Abstract
Chatbots powered by artificial intelligence have revolutionized many industries and fields of study, including medical education. Medical educators are increasingly asked to perform more administrative, written, and assessment functions with less time and resources. Safe use of chatbots, like ChatGPT, can help medical educators efficiently perform these functions. In this article, we provide medical educators with tips for the implementation of ChatGPT in medical education. Through creativity and careful construction of prompts, medical educators can use these and other implementations of chatbots, like ChatGPT, in their practice.
Collapse
Affiliation(s)
- Justin Peacock
- Department of Radiology and Radiological Sciences, Uniformed Services University, Bethesda, MD, USA
| | - Andrea Austin
- Department of Military and Emergency Medicine, Uniformed Services University, Bethesda, MD, USA
- UHS Southern California Education Consortium, Temecula, CA, USA
| | - Marina Shapiro
- Center for Health Professions Education, Uniformed Services University, Bethesda, MD, USA
| | - Alexis Battista
- Center for Health Professions Education, Uniformed Services University, Bethesda, MD, USA
| | - Anita Samuel
- Center for Health Professions Education, Uniformed Services University, Bethesda, MD, USA
| |
Collapse
|
25
|
Wei L. Artificial intelligence in language instruction: impact on English learning achievement, L2 motivation, and self-regulated learning. Front Psychol 2023; 14:1261955. [PMID: 38023040 PMCID: PMC10658009 DOI: 10.3389/fpsyg.2023.1261955] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2023] [Accepted: 10/11/2023] [Indexed: 12/01/2023] Open
Abstract
Introduction This mixed methods study examines the effects of AI-mediated language instruction on English learning achievement, L2 motivation, and self-regulated learning among English as a Foreign Language (EFL) learners. It addresses the increasing interest in AI-driven educational technologies and their potential to revolutionize language instruction. Methods Two intact classes, consisting of a total of 60 university students, participated in this study. The experimental group received AI-mediated instruction, while the control group received traditional language instruction. Pre-tests and post-tests were administered to evaluate English learning achievement across various domains, including grammar, vocabulary, reading comprehension, and writing skills. Additionally, self-report questionnaires were employed to assess L2 motivation and self-regulated learning. Results Quantitative analysis revealed that the experimental group achieved significantly higher English learning outcomes in all assessed areas compared to the control group. Furthermore, they exhibited greater L2 motivation and more extensive utilization of self-regulated learning strategies. These results suggest that AI-mediated instruction positively impacts English learning achievement, L2 motivation, and self-regulated learning. Discussion Qualitative analysis of semi-structured interviews with 14 students from the experimental group shed light on the transformative effects of the AI platform. It was found to enhance engagement and offer personalized learning experiences, ultimately boosting motivation and fostering self-regulated learning. These findings emphasize the potential of AI-mediated language instruction to improve language learning outcomes, motivate learners, and promote autonomy. Conclusion This study contributes to evidence-based language pedagogy, offering valuable insights to educators and researchers interested in incorporating AI-powered platforms into language classrooms. The results support the notion that AI-mediated language instruction holds promise in revolutionizing language learning, and it highlights the positive impact of AI-driven educational technologies in the realm of language education.
Collapse
Affiliation(s)
- Ling Wei
- College of Foreign Languages, Chongqing College of Mobile Communication, Chongqing, China
| |
Collapse
|
26
|
Abujaber AA, Abd-Alrazaq A, Al-Qudimat AR, Nashwan AJ. A Strengths, Weaknesses, Opportunities, and Threats (SWOT) Analysis of ChatGPT Integration in Nursing Education: A Narrative Review. Cureus 2023; 15:e48643. [PMID: 38090452 PMCID: PMC10711690 DOI: 10.7759/cureus.48643] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/11/2023] [Indexed: 03/25/2024] Open
Abstract
Amidst evolving healthcare demands, nursing education plays a pivotal role in preparing future nurses for complex challenges. Traditional approaches, however, must be revised to meet modern healthcare needs. The ChatGPT, an AI-based chatbot, has garnered significant attention due to its ability to personalize learning experiences, enhance virtual clinical simulations, and foster collaborative learning in nursing education. This review aims to thoroughly assess the potential impact of integrating ChatGPT into nursing education. The hypothesis is that valuable insights can be provided for stakeholders through a comprehensive SWOT analysis examining the strengths, weaknesses, opportunities, and threats associated with ChatGPT. This will enable informed decisions about its integration, prioritizing improved learning outcomes. A thorough narrative literature review was undertaken to provide a solid foundation for the SWOT analysis. The materials included scholarly articles and reports, which ensure the study's credibility and allow for a holistic and unbiased assessment. The analysis identified accessibility, consistency, adaptability, cost-effectiveness, and staying up-to-date as crucial factors influencing the strengths, weaknesses, opportunities, and threats associated with ChatGPT integration in nursing education. These themes provided a framework to understand the potential risks and benefits of integrating ChatGPT into nursing education. This review highlights the importance of responsible and effective use of ChatGPT in nursing education and the need for collaboration among educators, policymakers, and AI developers. Addressing the identified challenges and leveraging the strengths of ChatGPT can lead to improved learning outcomes and enriched educational experiences for students. The findings emphasize the importance of responsibly integrating ChatGPT in nursing education, balancing technological advancement with careful consideration of associated risks, to achieve optimal outcomes.
Collapse
Affiliation(s)
| | - Alaa Abd-Alrazaq
- AI Center for Precision Health, Weill Cornell Medicine-Qatar, Doha, QAT
| | - Ahmad R Al-Qudimat
- Department of Public Health, Qatar University, Doha, QAT
- Surgical Research Section, Department of Surgery, Hamad Medical Corporation, Doha, QAT
| | | |
Collapse
|
27
|
Lareyre F, Nasr B, Chaudhuri A, Di Lorenzo G, Carlier M, Raffort J. Comprehensive Review of Natural Language Processing (NLP) in Vascular Surgery. EJVES Vasc Forum 2023; 60:57-63. [PMID: 37822918 PMCID: PMC10562666 DOI: 10.1016/j.ejvsvf.2023.09.002] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2023] [Revised: 07/13/2023] [Accepted: 09/08/2023] [Indexed: 10/13/2023] Open
Abstract
Objective The use of Natural Language Processing (NLP) has attracted increased interest in healthcare with various potential applications including identification and extraction of health information, development of chatbots and virtual assistants. The aim of this comprehensive literature review was to provide an overview of NLP applications in vascular surgery, identify current limitations, and discuss future perspectives in the field. Data sources The MEDLINE database was searched on April 2023. Review methods The database was searched using a combination of keywords to identify studies reporting the use of NLP and chatbots in three main vascular diseases. Keywords used included Natural Language Processing, chatbot, chatGPT, aortic disease, carotid, peripheral artery disease, vascular, and vascular surgery. Results Given the heterogeneity of study design, techniques, and aims, a comprehensive literature review was performed to provide an overview of NLP applications in vascular surgery. By enabling identification and extraction of information on patients with vascular diseases, such technology could help to analyse data from healthcare information systems to provide feedback on current practice and help in optimising patient care. In addition, chatbots and NLP driven techniques have the potential to be used as virtual assistants for both health professionals and patients. Conclusion While Artificial Intelligence and NLP technology could be used to enhance care for patients with vascular diseases, many challenges remain including the need to define guidelines and clear consensus on how to evaluate and validate these innovations before their implementation into clinical practice.
Collapse
Affiliation(s)
- Fabien Lareyre
- Department of Vascular Surgery, Hospital of Antibes Juan-les-Pins, France
- Université Côte d'Azur, Inserm, U1065, C3M, Nice, France
| | - Bahaa Nasr
- Department of Vascular and Endovascular Surgery, Brest University Hospital, Brest, France
- INSERM, UMR 1101, LaTIM, Brest, France
| | - Arindam Chaudhuri
- Bedfordshire - Milton Keynes Vascular Centre, Bedfordshire Hospitals, NHS Foundation Trust, Bedford, UK
| | - Gilles Di Lorenzo
- Department of Vascular Surgery, Hospital of Antibes Juan-les-Pins, France
| | - Mathieu Carlier
- Department of Urology, University Hospital of Nice, Nice, France
| | - Juliette Raffort
- Université Côte d'Azur, Inserm, U1065, C3M, Nice, France
- Institute 3IA Côte d’Azur, Université Côte d’Azur, France
- Clinical Chemistry Laboratory, University Hospital of Nice, France
| |
Collapse
|
28
|
Garg RK, Urs VL, Agarwal AA, Chaudhary SK, Paliwal V, Kar SK. Exploring the role of ChatGPT in patient care (diagnosis and treatment) and medical research: A systematic review. Health Promot Perspect 2023; 13:183-191. [PMID: 37808939 PMCID: PMC10558973 DOI: 10.34172/hpp.2023.22] [Citation(s) in RCA: 58] [Impact Index Per Article: 29.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 07/06/2023] [Indexed: 10/10/2023] Open
Abstract
BACKGROUND ChatGPT is an artificial intelligence based tool developed by OpenAI (California, USA). This systematic review examines the potential of ChatGPT in patient care and its role in medical research. METHODS The systematic review was done according to the PRISMA guidelines. Embase, Scopus, PubMed and Google Scholar data bases were searched. We also searched preprint data bases. Our search was aimed to identify all kinds of publications, without any restrictions, on ChatGPT and its application in medical research, medical publishing and patient care. We used search term "ChatGPT". We reviewed all kinds of publications including original articles, reviews, editorial/ commentaries, and even letter to the editor. Each selected records were analysed using ChatGPT and responses generated were compiled in a table. The word table was transformed in to a PDF and was further analysed using ChatPDF. RESULTS We reviewed full texts of 118 articles. ChatGPT can assist with patient enquiries, note writing, decision-making, trial enrolment, data management, decision support, research support, and patient education. But the solutions it offers are usually insufficient and contradictory, raising questions about their originality, privacy, correctness, bias, and legality. Due to its lack of human-like qualities, ChatGPT's legitimacy as an author is questioned when used for academic writing. ChatGPT generated contents have concerns with bias and possible plagiarism. CONCLUSION Although it can help with patient treatment and research, there are issues with accuracy, authorship, and bias. ChatGPT can serve as a "clinical assistant" and be a help in research and scholarly writing.
Collapse
Affiliation(s)
| | - Vijeth L Urs
- Department of Neurology, King George’s Medical University, Lucknow, India
| | | | | | - Vimal Paliwal
- Department of Neurology, Sanjay Gandhi Institute of Medical Sciences, Lucknow, India
| | - Sujita Kumar Kar
- Department of Psychiatry, King George’s Medical University, Lucknow, India
| |
Collapse
|
29
|
Walters WH, Wilder EI. Fabrication and errors in the bibliographic citations generated by ChatGPT. Sci Rep 2023; 13:14045. [PMID: 37679503 PMCID: PMC10484980 DOI: 10.1038/s41598-023-41032-5] [Citation(s) in RCA: 55] [Impact Index Per Article: 27.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Accepted: 08/21/2023] [Indexed: 09/09/2023] Open
Abstract
Although chatbots such as ChatGPT can facilitate cost-effective text generation and editing, factually incorrect responses (hallucinations) limit their utility. This study evaluates one particular type of hallucination: fabricated bibliographic citations that do not represent actual scholarly works. We used ChatGPT-3.5 and ChatGPT-4 to produce short literature reviews on 42 multidisciplinary topics, compiling data on the 636 bibliographic citations (references) found in the 84 papers. We then searched multiple databases and websites to determine the prevalence of fabricated citations, to identify errors in the citations to non-fabricated papers, and to evaluate adherence to APA citation format. Within this set of documents, 55% of the GPT-3.5 citations but just 18% of the GPT-4 citations are fabricated. Likewise, 43% of the real (non-fabricated) GPT-3.5 citations but just 24% of the real GPT-4 citations include substantive citation errors. Although GPT-4 is a major improvement over GPT-3.5, problems remain.
Collapse
Affiliation(s)
- William H Walters
- Mary Alice & Tom O'Malley Library, Manhattan College, Riverdale, NY, USA.
| | - Esther Isabelle Wilder
- Department of Sociology, Lehman College, The City University of New York, Bronx, NY, USA
- Doctoral Program in Sociology, CUNY Graduate Center, The City University of New York, New York, NY, USA
| |
Collapse
|
30
|
Talyshinskii A, Naik N, Hameed BMZ, Zhanbyrbekuly U, Khairli G, Guliev B, Juilebø-Jones P, Tzelves L, Somani BK. Expanding horizons and navigating challenges for enhanced clinical workflows: ChatGPT in urology. Front Surg 2023; 10:1257191. [PMID: 37744723 PMCID: PMC10512827 DOI: 10.3389/fsurg.2023.1257191] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2023] [Accepted: 08/28/2023] [Indexed: 09/26/2023] Open
Abstract
Purpose of review ChatGPT has emerged as a potential tool for facilitating doctors' workflows. However, when it comes to applying these findings within a urological context, there have not been many studies. Thus, our objective was rooted in analyzing the pros and cons of ChatGPT use and how it can be exploited and used by urologists. Recent findings ChatGPT can facilitate clinical documentation and note-taking, patient communication and support, medical education, and research. In urology, it was proven that ChatGPT has the potential as a virtual healthcare aide for benign prostatic hyperplasia, an educational and prevention tool on prostate cancer, educational support for urological residents, and as an assistant in writing urological papers and academic work. However, several concerns about its exploitation are presented, such as lack of web crawling, risk of accidental plagiarism, and concerns about patients-data privacy. Summary The existing limitations mediate the need for further improvement of ChatGPT, such as ensuring the privacy of patient data and expanding the learning dataset to include medical databases, and developing guidance on its appropriate use. Urologists can also help by conducting studies to determine the effectiveness of ChatGPT in urology in clinical scenarios and nosologies other than those previously listed.
Collapse
Affiliation(s)
- Ali Talyshinskii
- Department of Urology, Astana Medical University, Astana, Kazakhstan
| | - Nithesh Naik
- Department of Mechanical and Industrial Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, India
| | | | | | - Gafur Khairli
- Department of Urology, Astana Medical University, Astana, Kazakhstan
| | - Bakhman Guliev
- Department of Urology, Mariinsky Hospital, St Petersburg, Russia
| | | | - Lazaros Tzelves
- Department of Urology, National and Kapodistrian University of Athens, Sismanogleion Hospital, Athens, Marousi, Greece
| | - Bhaskar Kumar Somani
- Department of Urology, University Hospital Southampton NHS Trust, Southampton, United Kingdom
| |
Collapse
|
31
|
Watters C, Lemanski MK. Universal skepticism of ChatGPT: a review of early literature on chat generative pre-trained transformer. Front Big Data 2023; 6:1224976. [PMID: 37680954 PMCID: PMC10482048 DOI: 10.3389/fdata.2023.1224976] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 07/10/2023] [Indexed: 09/09/2023] Open
Abstract
ChatGPT, a new language model developed by OpenAI, has garnered significant attention in various fields since its release. This literature review provides an overview of early ChatGPT literature across multiple disciplines, exploring its applications, limitations, and ethical considerations. The review encompasses Scopus-indexed publications from November 2022 to April 2023 and includes 156 articles related to ChatGPT. The findings reveal a predominance of negative sentiment across disciplines, though subject-specific attitudes must be considered. The review highlights the implications of ChatGPT in many fields including healthcare, raising concerns about employment opportunities and ethical considerations. While ChatGPT holds promise for improved communication, further research is needed to address its capabilities and limitations. This literature review provides insights into early research on ChatGPT, informing future investigations and practical applications of chatbot technology, as well as development and usage of generative AI.
Collapse
Affiliation(s)
- Casey Watters
- Faculty of Law, Bond University, Gold Coast, QLD, Australia
| | | |
Collapse
|
32
|
Hsu HY, Hsu KC, Hou SY, Wu CL, Hsieh YW, Cheng YD. Examining Real-World Medication Consultations and Drug-Herb Interactions: ChatGPT Performance Evaluation. JMIR MEDICAL EDUCATION 2023; 9:e48433. [PMID: 37561097 PMCID: PMC10477918 DOI: 10.2196/48433] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Revised: 06/23/2023] [Accepted: 07/25/2023] [Indexed: 08/11/2023]
Abstract
BACKGROUND Since OpenAI released ChatGPT, with its strong capability in handling natural tasks and its user-friendly interface, it has garnered significant attention. OBJECTIVE A prospective analysis is required to evaluate the accuracy and appropriateness of medication consultation responses generated by ChatGPT. METHODS A prospective cross-sectional study was conducted by the pharmacy department of a medical center in Taiwan. The test data set comprised retrospective medication consultation questions collected from February 1, 2023, to February 28, 2023, along with common questions about drug-herb interactions. Two distinct sets of questions were tested: real-world medication consultation questions and common questions about interactions between traditional Chinese and Western medicines. We used the conventional double-review mechanism. The appropriateness of each response from ChatGPT was assessed by 2 experienced pharmacists. In the event of a discrepancy between the assessments, a third pharmacist stepped in to make the final decision. RESULTS Of 293 real-world medication consultation questions, a random selection of 80 was used to evaluate ChatGPT's performance. ChatGPT exhibited a higher appropriateness rate in responding to public medication consultation questions compared to those asked by health care providers in a hospital setting (31/51, 61% vs 20/51, 39%; P=.01). CONCLUSIONS The findings from this study suggest that ChatGPT could potentially be used for answering basic medication consultation questions. Our analysis of the erroneous information allowed us to identify potential medical risks associated with certain questions; this problem deserves our close attention.
Collapse
Affiliation(s)
- Hsing-Yu Hsu
- Department of Pharmacy, China Medical University Hospital, Taichung, Taiwan
- Graduate Institute of Clinical Pharmacy, College of Medicine, National Taiwan University, Taipei, Taiwan
| | - Kai-Cheng Hsu
- Artificial Intelligence Center, China Medical University Hospital, Taichung, Taiwan
- Department of Medicine, China Medical University, Taichung, Taiwan
| | - Shih-Yen Hou
- Artificial Intelligence Center, China Medical University Hospital, Taichung, Taiwan
| | - Ching-Lung Wu
- School of Pharmacy, College of Pharmacy, China Medical University, Taichung, Taiwan
| | - Yow-Wen Hsieh
- Department of Pharmacy, China Medical University Hospital, Taichung, Taiwan
- School of Pharmacy, College of Pharmacy, China Medical University, Taichung, Taiwan
| | - Yih-Dih Cheng
- Department of Pharmacy, China Medical University Hospital, Taichung, Taiwan
- School of Pharmacy, College of Pharmacy, China Medical University, Taichung, Taiwan
| |
Collapse
|
33
|
Sharma SC, Ramchandani JP, Thakker A, Lahiri A. ChatGPT in Plastic and Reconstructive Surgery. Indian J Plast Surg 2023; 56:320-325. [PMID: 37705820 PMCID: PMC10497341 DOI: 10.1055/s-0043-1771514] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/15/2023] Open
Abstract
Background Chat Generative Pre-Trained Transformer (ChatGPT) is a versatile large language model-based generative artificial intelligence. It is proficient in a variety of tasks from drafting emails to coding to composing music to passing medical licensing exams. While the potential role of ChatGPT in plastic surgery is promising, evidence-based research is needed to guide its implementation in practice. Methods This review aims to summarize the literature surrounding ChatGPT's use in plastic surgery. Results A literature search revealed several applications for ChatGPT in the field of plastic surgery, including the ability to create academic literature and to aid the production of research. However, the ethical implications of using such chatbots in scientific writing requires careful consideration. ChatGPT can also generate high-quality patient discharge summaries and operation notes within seconds, freeing up busy junior doctors to complete other tasks. However, currently clinical information must still be manually inputted, and clinicians must consider data privacy implications. Its use in aiding patient communication and education and training is also widely documented in the literature. However, questions have been raised over the accuracy of answers generated given that current versions of ChatGPT cannot access the most up-to-date sources. Conclusions While one must be aware of its shortcomings, ChatGPT is a useful tool for plastic surgeons to improve productivity for a range of tasks from manuscript preparation to healthcare communication generation to drafting teaching sessions to studying and learning. As access improves and technology becomes more refined, surely more uses for ChatGPT in plastic surgery will become apparent.
Collapse
Affiliation(s)
- Sanjeev Chaand Sharma
- Department of Plastic Surgery, Leicester Royal Infirmary, Infirmary Square, Leicester, United Kingdom
| | - Jai Parkash Ramchandani
- Faculty of Life Sciences & Medicine, King's College London, Guy's Campus, Great Maze Pond, London, United Kingdom
| | - Arjuna Thakker
- Academic Team of Musculoskeletal Surgery, Leicester General Hospital, University Hospitals of Leicester NHS Trust, United Kingdom
| | - Anindya Lahiri
- Department of Plastic Surgery, Sandwell General Hospital, West Bromwich, United Kingdom
| |
Collapse
|
34
|
Mago J, Sharma M. The Potential Usefulness of ChatGPT in Oral and Maxillofacial Radiology. Cureus 2023; 15:e42133. [PMID: 37476297 PMCID: PMC10355343 DOI: 10.7759/cureus.42133] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/19/2023] [Indexed: 07/22/2023] Open
Abstract
Aim This study aimed to evaluate the potential usefulness of Chat Generated Pre-Trained Transformer-3 (ChatGPT-3) in oral and maxillofacial radiology for report writing by identifying radiographic anatomical landmarks and learning about oral and maxillofacial pathologies and their radiographic features. The study also aimed to evaluate the performance of ChatGPT-3 and its usage in oral and maxillofacial radiology training. Materials and methods A questionnaire consisting of 80 questions was queried on the OpenAI app ChatGPT-3. The questions were stratified based on three categories. The categorization was based on random anatomical landmarks, oral and maxillofacial pathologies, and the radiographic features of some of these pathologies. One oral and maxillofacial radiologist evaluated queries that were answered by the ChatGPT-3 model and rated them on a 4-point, modified Likert scale. The post-survey analysis for the performance of ChatGPT-3 was based on the Strengths, Weaknesses, Opportunities, and Threats (SWOT) analysis, its application in oral and maxillofacial radiology training, and its recommended use. Results In order of efficiency, Chat GPT-3 gave 100% accuracy in describing radiographic landmarks. However, the content of the oral and maxillofacial pathologies was limited to major or characteristic radiographic features. The mean scores for the queries related to the anatomic landmarks, oral and maxillofacial pathologies, and radiographic features of the oral and maxillofacial pathologies were 3.94, 3.85, and 3.96, respectively. However, the median and mode scores were 4 and were similar to all categories. The data for the oral and maxillofacial pathologies when the questions were not specifically included in the format of the introduction of the pathology, causes, symptoms, and treatment. Out of two abbreviations, one was not answered correctly. Conclusion The study showed that ChatGPT-3 is efficient in describing the pathology, characteristic radiographic features, and describing anatomical landmarks. ChatGPT-3 can be used as an adjunct when an oral radiologist needs additional information on any pathology, however, it cannot be the mainstay for reference. ChatGPT-3 is less detail-oriented, and the data has a risk of infodemics and the possibility of medical errors. However, Chat GPT-3 can be an excellent tool in helping the community in increasing the knowledge and awareness of various pathologies and decreasing the anxiety of the patients while dental healthcare professionals formulate an appropriate treatment plan.
Collapse
Affiliation(s)
- Jyoti Mago
- Oral and Maxillofacial Radiology, University of Nevada, Las Vegas (UNLV), Las Vegas, USA
| | - Manoj Sharma
- Public Health, University of Nevada, Las Vegas (UNLV), Las Vegas, USA
| |
Collapse
|
35
|
Sifat RI, Bhattacharya U. Transformative potential of artificial intelligence in global health policy. JOURNAL OF MARKET ACCESS & HEALTH POLICY 2023; 11:2230660. [PMID: 37405227 PMCID: PMC10316731 DOI: 10.1080/20016689.2023.2230660] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 07/06/2023]
Affiliation(s)
| | - Upali Bhattacharya
- Department of Sociology, Virginia Polytechnic Institute and State University, Blacksburg, VA, USA
| |
Collapse
|
36
|
Ocampo TSC, Silva TP, Alencar-Palha C, Haiter-Neto F, Oliveira ML. ChatGPT and scientific writing: A reflection on the ethical boundaries. Imaging Sci Dent 2023; 53:175-176. [PMID: 37405199 PMCID: PMC10315235 DOI: 10.5624/isd.20230085] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2023] [Revised: 04/20/2023] [Accepted: 05/03/2023] [Indexed: 07/06/2023] Open
Affiliation(s)
- Thaís Santos Cerqueira Ocampo
- Division of Oral Radiology, Department of Oral Diagnosis, Piracicaba Dental School, University of Campinas, Piracicaba, SP, Brazil
| | - Thaísa Pinheiro Silva
- Division of Oral Radiology, Department of Oral Diagnosis, Piracicaba Dental School, University of Campinas, Piracicaba, SP, Brazil
| | - Caio Alencar-Palha
- Division of Oral Radiology, Department of Oral Diagnosis, Piracicaba Dental School, University of Campinas, Piracicaba, SP, Brazil
| | - Francisco Haiter-Neto
- Division of Oral Radiology, Department of Oral Diagnosis, Piracicaba Dental School, University of Campinas, Piracicaba, SP, Brazil
| | - Matheus L. Oliveira
- Division of Oral Radiology, Department of Oral Diagnosis, Piracicaba Dental School, University of Campinas, Piracicaba, SP, Brazil
| |
Collapse
|
37
|
Komorowski M, Del Pilar Arias López M, Chang AC. How could ChatGPT impact my practice as an intensivist? An overview of potential applications, risks and limitations. Intensive Care Med 2023:10.1007/s00134-023-07096-7. [PMID: 37256340 DOI: 10.1007/s00134-023-07096-7] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2023] [Accepted: 05/05/2023] [Indexed: 06/01/2023]
Affiliation(s)
- Matthieu Komorowski
- Division of Anaesthetics, Pain Medicine, and Intensive Care, Department of Surgery and Cancer, Faculty of Medicine, Imperial College London, London, SW7 2AZ, UK.
- Intensive Care Unit, Charing Cross Hospital, Fulham Palace Road, London, W6 8RF, UK.
| | - Maria Del Pilar Arias López
- Hospital de Niños Ricardo Gutierrez, Intermediate Care Unit, Gallo 1330, C1425EFD, Buenos Aires, Argentina
- Argentina Society of Intensive Care, SATI-Q Paediatric Program. Av. Cnel, Niceto Vega 4617, C1414BEA, Buenos Aires, Argentina
| | - Anthony C Chang
- Children's Hospital of Orange County Sharon Disney Lund Medical Intelligence and Innovation Institute, 1201 W. La Veta Ave, Orange, CA, 92868-3874, USA
| |
Collapse
|
38
|
Seetharaman R. Revolutionizing Medical Education: Can ChatGPT Boost Subjective Learning and Expression? J Med Syst 2023; 47:61. [PMID: 37160568 DOI: 10.1007/s10916-023-01957-w] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2023] [Accepted: 04/20/2023] [Indexed: 05/11/2023]
Abstract
ChatGPT is an AI tool that can be used to enhance medical education by helping students develop subjective learning and expression skills. These skills are critical in clinical practice, but the current medical education system is heavily focused on objective assessments, such as Multiple Choice Questions (MCQs). Students from non-English speaking backgrounds can particularly struggle with expressing themselves in English, which is the primary language of instruction in many medical schools worldwide. ChatGPT can provide additional language support for these students to help them develop their language skills and communicate effectively. ChatGPT can be used in small group assessments to serve as a benchmark for students to strive for in their medical education. By comparing their answers to ChatGPT's responses, students can identify gaps in their knowledge and work to fill them. ChatGPT can also provide students with feedback on their writing style and language usage, helping them to improve their subjective expression of medical knowledge. Furthermore, ChatGPT can be used to simulate patient encounters for medical students. By interacting with ChatGPT, students can practice taking medical histories and documenting symptoms accurately. In continuing medical education (CME) programs, physicians can also benefit from ChatGPT's capabilities. By using ChatGPT to search for the latest research articles, clinical trials, and treatment guidelines, physicians can stay informed and provide the best care possible to their patients. Overall, ChatGPT has the potential to be a valuable tool in medical education by helping students and physicians develop the essential skills required for clinical practice, such as communication, problem-solving, and critical thinking.
Collapse
Affiliation(s)
- Rajmohan Seetharaman
- Department of Pharmacology and Therapeutics, Seth G.S. Medical College & KEM Hospital, Parel, Mumbai, 400012, India.
| |
Collapse
|
39
|
Heck TG. What artificial intelligence knows about 70 kDa heat shock proteins, and how we will face this ChatGPT era. Cell Stress Chaperones 2023; 28:225-229. [PMID: 37058213 PMCID: PMC10103022 DOI: 10.1007/s12192-023-01340-1] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Revised: 03/17/2023] [Accepted: 03/24/2023] [Indexed: 04/15/2023] Open
Affiliation(s)
- Thiago Gomes Heck
- Post Graduate Program in Integral Health Care (PPGAIS-UNIJUÍ/UNICRUZ/URI), Regional University of Northwestern Rio Grande Do Sul State (UNIJUI), Ijuí, RS, Brazil.
- Post Graduate Program in Mathematical and Computational Modeling (PPGMMC), Regional University of Northwestern Rio Grande Do Sul State (UNIJUI), Ijuí, RS, Brazil.
| |
Collapse
|
40
|
Sifat RI. ChatGPT and the Future of Health Policy Analysis: Potential and Pitfalls of Using ChatGPT in Policymaking. Ann Biomed Eng 2023:10.1007/s10439-023-03204-2. [PMID: 37061595 DOI: 10.1007/s10439-023-03204-2] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2023] [Accepted: 04/11/2023] [Indexed: 04/17/2023]
Abstract
Scholars increasingly rely on new artificial intelligence models for convenience and simple access to necessities due to the rapid evolution of scientific literature and technology. The invention of ChatGPT by OpenAI stands out as a key example of how significant advances in large language model technology have recently changed the field of artificial intelligence (AI). Since ChatGPT's development, it has been tested by multiple sectors on various topics to see how well it functions in a natural and conversational mode. The crucial question is how much ChatGPT can influence global health policy analysis. In this article, the researcher briefly explains ChatGPT's potential and the difficulties that users, such as researchers or policymakers, may continue to face.
Collapse
Affiliation(s)
- Ridwan Islam Sifat
- School of Public Policy, University of Maryland, Baltimore County, Baltimore, MD, 21250, USA.
| |
Collapse
|
41
|
Alhaidry HM, Fatani B, Alrayes JO, Almana AM, Alfhaed NK. ChatGPT in Dentistry: A Comprehensive Review. Cureus 2023; 15:e38317. [PMID: 37266053 PMCID: PMC10230850 DOI: 10.7759/cureus.38317] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/29/2023] [Indexed: 06/03/2023] Open
Abstract
Chat generative pre-trained transformer (ChatGPT) is an artificial intelligence chatbot that uses natural language processing that can respond to human input in a conversational manner. ChatGPT has numerous applications in the health care system including dentistry; it is used in diagnoses and for assessing disease risk and scheduling appointments. It also has a role in scientific research. In the dental field, it has provided many benefits such as detecting dental and maxillofacial abnormalities on panoramic radiographs and identifying different dental restorations. Therefore, it helps in decreasing the workload. But even with these benefits, one should take into consideration the risks and limitations of this chatbot. Few articles mentioned the use of ChatGPT in dentistry. This comprehensive review represents data collected from 66 relevant articles using PubMed and Google Scholar as databases. This review aims to discuss all relevant published articles on the use of ChatGPT in dentistry.
Collapse
Affiliation(s)
- Hind M Alhaidry
- Advanced General Dentistry, Prince Sultan Military Medical City, Riyadh, SAU
| | - Bader Fatani
- Dentistry, College of Dentistry, King Saud University, Riyadh, SAU
| | - Jenan O Alrayes
- Dentistry, College of Dentistry, King Saud University, Riyadh, SAU
| | | | - Nawaf K Alfhaed
- Dentistry, College of Dentistry, King Saud University, Riyadh, SAU
| |
Collapse
|