101
|
Ganjavi C, Eppler MB, Pekcan A, Biedermann B, Abreu A, Collins GS, Gill IS, Cacciamani GE. Publishers' and journals' instructions to authors on use of generative artificial intelligence in academic and scientific publishing: bibliometric analysis. BMJ 2024; 384:e077192. [PMID: 38296328 PMCID: PMC10828852 DOI: 10.1136/bmj-2023-077192] [Citation(s) in RCA: 28] [Impact Index Per Article: 28.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 11/29/2023] [Indexed: 02/05/2024]
Abstract
OBJECTIVES To determine the extent and content of academic publishers' and scientific journals' guidance for authors on the use of generative artificial intelligence (GAI). DESIGN Cross sectional, bibliometric study. SETTING Websites of academic publishers and scientific journals, screened on 19-20 May 2023, with the search updated on 8-9 October 2023. PARTICIPANTS Top 100 largest academic publishers and top 100 highly ranked scientific journals, regardless of subject, language, or country of origin. Publishers were identified by the total number of journals in their portfolio, and journals were identified through the Scimago journal rank using the Hirsch index (H index) as an indicator of journal productivity and impact. MAIN OUTCOME MEASURES The primary outcomes were the content of GAI guidelines listed on the websites of the top 100 academic publishers and scientific journals, and the consistency of guidance between the publishers and their affiliated journals. RESULTS Among the top 100 largest publishers, 24% provided guidance on the use of GAI, of which 15 (63%) were among the top 25 publishers. Among the top 100 highly ranked journals, 87% provided guidance on GAI. Of the publishers and journals with guidelines, the inclusion of GAI as an author was prohibited in 96% and 98%, respectively. Only one journal (1%) explicitly prohibited the use of GAI in the generation of a manuscript, and two (8%) publishers and 19 (22%) journals indicated that their guidelines exclusively applied to the writing process. When disclosing the use of GAI, 75% of publishers and 43% of journals included specific disclosure criteria. Where to disclose the use of GAI varied, including in the methods or acknowledgments, in the cover letter, or in a new section. Variability was also found in how to access GAI guidelines shared between journals and publishers. GAI guidelines in 12 journals directly conflicted with those developed by the publishers. The guidelines developed by top medical journals were broadly similar to those of academic journals. CONCLUSIONS Guidelines by some top publishers and journals on the use of GAI by authors are lacking. Among those that provided guidelines, the allowable uses of GAI and how it should be disclosed varied substantially, with this heterogeneity persisting in some instances among affiliated publishers and journals. Lack of standardization places a burden on authors and could limit the effectiveness of the regulations. As GAI continues to grow in popularity, standardized guidelines to protect the integrity of scientific output are needed.
Collapse
Affiliation(s)
- Conner Ganjavi
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Michael B Eppler
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Asli Pekcan
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Brett Biedermann
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Andre Abreu
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Gary S Collins
- UK EQUATOR Centre, Centre for Statistics in Medicine, Nuffield Department of Orthopaedics, Rheumatology and Musculoskeletal Sciences, University of Oxford, Oxford, UK
| | - Inderbir S Gill
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| | - Giovanni E Cacciamani
- Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
102
|
Islam N, van der Schaar M. Use of generative artificial intelligence in medical research. BMJ 2024; 384:q119. [PMID: 38296355 DOI: 10.1136/bmj.q119] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 02/08/2024]
Affiliation(s)
- Nazrul Islam
- Faculty of Medicine, University of Southampton, Southampton, UK
| | | |
Collapse
|
103
|
Jairoun AA, El-Dahiyat F, ElRefae GA, Al-Hemyari SS, Shahwan M, Zyoud SH, Hammour KA, Babar ZUD. Detecting manuscripts written by generative AI and AI-assisted technologies in the field of pharmacy practice. J Pharm Policy Pract 2024; 17:2303759. [PMID: 38229951 PMCID: PMC10791078 DOI: 10.1080/20523211.2024.2303759] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/18/2024] Open
Abstract
Generative AI can be a powerful research tool, but researchers must employ it ethically and transparently. This commentary addresses how the editors of pharmacy practice journals can identify manuscripts generated by generative AI and AI-assisted technologies. Editors and reviewers must stay well-informed about developments in AI technologies to effectively recognise AI-written papers. Editors should safeguard the reliability of journal publishing and sustain industry standards for pharmacy practice by implementing the crucial strategies outlined in this editorial. Although obstacles, including ignorance, time constraints, and protean AI strategies, might hinder detection efforts, several facilitators can help overcome those obstacles. Pharmacy practice journal editors and reviewers would benefit from educational programmes, collaborations with AI experts, and sophisticated plagiarism-detection techniques geared toward accurately identifying AI-generated text. Academics and practitioners can further uphold the integrity of published research through transparent reporting and ethical standards. Pharmacy practice journal staffs can sustain academic rigour and guarantee the validity of scholarly work by recognising and addressing the relevant barriers and utilising the proper enablers. Navigating the changing world of AI-generated content and preserving standards of excellence in pharmaceutical research and practice requires a proactive strategy of constant learning and community participation.
Collapse
Affiliation(s)
- Ammar Abdulrahman Jairoun
- Health and Safety Department, Dubai Municipality, Dubai, UAE
- Discipline of Clinical Pharmacy, School of Pharmaceutical Sciences, Universiti Sains Malaysia (USM), George Town, Malaysia
| | - Faris El-Dahiyat
- Clinical Pharmacy Program, College of Pharmacy, Al Ain University, Al Ain, UAE
- Artificial Intelligence Research Center, Al Ain University, Al Ain, UAE
| | - Ghaleb A. ElRefae
- Artificial Intelligence Research Center, Al Ain University, Al Ain, UAE
| | - Sabaa Saleh Al-Hemyari
- Discipline of Clinical Pharmacy, School of Pharmaceutical Sciences, Universiti Sains Malaysia (USM), George Town, Malaysia
- Pharmacy Department, Emirates Health Services, Dubai, UAE
| | - Moyad Shahwan
- Centre of Medical and Bio-allied Health Sciences Research, Ajman University, Ajman, UAE
- Department of Clinical Sciences, College of Pharmacy and Health Sciences, Ajman University, Ajman, UAE
| | - Samer H. Zyoud
- Department of Mathematics and Sciences, Ajman University, Ajman, UAE
| | - Khawla Abu Hammour
- Department of Biopharmaceutics and Clinical Pharmacy, Faculty of Pharmacy, The University of Jordan, Amman, Jordan
| | - Zaheer-Ud-Din Babar
- Department of Pharmacy, School of Applied Sciences, University of Huddersfield, Huddersfield, UK
| |
Collapse
|
104
|
Farhat F, Silva ES, Hassani H, Madsen DØ, Sohail SS, Himeur Y, Alam MA, Zafar A. The scholarly footprint of ChatGPT: a bibliometric analysis of the early outbreak phase. Front Artif Intell 2024; 6:1270749. [PMID: 38249789 PMCID: PMC10797012 DOI: 10.3389/frai.2023.1270749] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Accepted: 12/08/2023] [Indexed: 01/23/2024] Open
Abstract
This paper presents a comprehensive analysis of the scholarly footprint of ChatGPT, an AI language model, using bibliometric and scientometric methods. The study zooms in on the early outbreak phase from when ChatGPT was launched in November 2022 to early June 2023. It aims to understand the evolution of research output, citation patterns, collaborative networks, application domains, and future research directions related to ChatGPT. By retrieving data from the Scopus database, 533 relevant articles were identified for analysis. The findings reveal the prominent publication venues, influential authors, and countries contributing to ChatGPT research. Collaborative networks among researchers and institutions are visualized, highlighting patterns of co-authorship. The application domains of ChatGPT, such as customer support and content generation, are examined. Moreover, the study identifies emerging keywords and potential research areas for future exploration. The methodology employed includes data extraction, bibliometric analysis using various indicators, and visualization techniques such as Sankey diagrams. The analysis provides valuable insights into ChatGPT's early footprint in academia and offers researchers guidance for further advancements. This study stimulates discussions, collaborations, and innovations to enhance ChatGPT's capabilities and impact across domains.
Collapse
Affiliation(s)
- Faiza Farhat
- Department of Zoology, Aligarh Muslim University, Aligarh, India
| | - Emmanuel Sirimal Silva
- Department of Economics and Law, Glasgow School for Business and Society, Glasgow Caledonian University, Glasgow, United Kingdom
| | - Hossein Hassani
- The Research Institute of Energy Management and Planning (RIEMP), University of Tehran, Tehran, Iran
| | - Dag Øivind Madsen
- USN School of Business, University of South-Eastern Norway, Hønefoss, Norway
| | - Shahab Saquib Sohail
- Department of Computer Science and Engineering, School of Engineering Sciences and Technology, Jamia Hamdard, New Delhi, India
| | - Yassine Himeur
- College of Engineering and Information Technology, University of Dubai, Dubai, United Arab Emirates
| | - M. Afshar Alam
- Department of Computer Science and Engineering, School of Engineering Sciences and Technology, Jamia Hamdard, New Delhi, India
| | - Aasim Zafar
- Department of Computer Science, Aligarh Muslim University, Aligarh, India
| |
Collapse
|
105
|
Ting DSJ, Tan TF, Ting DSW. ChatGPT in ophthalmology: the dawn of a new era? Eye (Lond) 2024; 38:4-7. [PMID: 37369764 PMCID: PMC10764795 DOI: 10.1038/s41433-023-02619-4] [Citation(s) in RCA: 26] [Impact Index Per Article: 26.0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2023] [Revised: 05/22/2023] [Accepted: 06/02/2023] [Indexed: 06/29/2023] Open
Affiliation(s)
- Darren Shu Jeng Ting
- Birmingham and Midland Eye Centre, Birmingham, UK
- Academic Unit of Ophthalmology, Institute of Inflammation and Ageing, University of Birmingham, Birmingham, UK
- Academic Ophthalmology, School of Medicine, University of Nottingham, Nottingham, UK
| | - Ting Fang Tan
- Artificial Intelligence and Digital Innovation Research Group, Singapore Eye Research Institute, Singapore, Singapore
- Singapore National Eye Centre, Singapore, Singapore
| | - Daniel Shu Wei Ting
- Artificial Intelligence and Digital Innovation Research Group, Singapore Eye Research Institute, Singapore, Singapore.
- Singapore National Eye Centre, Singapore, Singapore.
- Department of Ophthalmology and Visual Sciences, Duke-National University of Singapore Medical School, Singapore, Singapore.
| |
Collapse
|
106
|
Di Ieva A, Stewart C, Suero Molina E. Large Language Models in Neurosurgery. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2024; 1462:177-198. [PMID: 39523266 DOI: 10.1007/978-3-031-64892-2_11] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2024]
Abstract
A large language model (LLM), in the context of natural language processing and artificial intelligence, refers to a sophisticated neural network that has been trained on a massive amount of text data to understand and generate human-like language. These models are typically built on architectures like transformers. The term "large" indicates that the neural network has a significant number of parameters, making it more powerful and capable of capturing complex patterns in language. One notable example of a large language model is ChatGPT. ChatGPT is a large language model developed by OpenAI that uses deep learning techniques to generate human-like text. It can be trained on a variety of tasks, such as language translation, question answering, and text completion. One of the key features of ChatGPT is its ability to understand and respond to natural language inputs. This makes it a powerful tool for generating a wide range of text, including medical reports, surgical notes, and even poetry. Additionally, the model has been trained on a large corpus of text, which allows it to generate text that is both grammatically correct and semantically meaningful. In terms of applications in neurosurgery, ChatGPT can be used to generate detailed and accurate surgical reports, which can be very useful for sharing information about a patient's case with other members of the medical team. Additionally, the model can be used to generate detailed surgical notes, which can be very useful for training and educating residents and medical students. Overall, LLMs have the potential to be a valuable tool in the field of neurosurgery. Indeed, this abstract has been generated by ChatGPT within few seconds. Potential applications and pitfalls of the applications of LLMs are discussed in this paper.
Collapse
Affiliation(s)
- Antonio Di Ieva
- Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Faculty of Medicine, Human and Health Sciences, Macquarie University, Sydney, NSW, Australia.
- Macquarie Neurosurgery & Spine, MQ Health, Macquarie University Hospital, Sydney, NSW, Australia.
- Department of Neurosurgery, Nepean Blue Mountains Local Health District, Penrith, NSW, Australia.
- Centre for Applied Artificial Intelligence, School of Computing, Macquarie University, Sydney, NSW, Australia.
| | - Caleb Stewart
- Department of Neurosurgery, Louisiana State University Health Sciences Shreveport, Shreveport, LA, USA
| | - Eric Suero Molina
- Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Faculty of Medicine, Human and Health Sciences, Macquarie University, Sydney, NSW, Australia
- Department of Neurosurgery, University Hospital of Münster, Münster, Germany
| |
Collapse
|
107
|
Warren E, Hurley ET, Park CN, Crook BS, Lorentz S, Levin JM, Anakwenze O, MacDonald PB, Klifto CS. Evaluation of information from artificial intelligence on rotator cuff repair surgery. JSES Int 2024; 8:53-57. [PMID: 38312282 PMCID: PMC10837709 DOI: 10.1016/j.jseint.2023.09.009] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/06/2024] Open
Abstract
Purpose The purpose of this study was to analyze the quality and readability of information regarding rotator cuff repair surgery available using an online AI software. Methods An open AI model (ChatGPT) was used to answer 24 commonly asked questions from patients on rotator cuff repair. Questions were stratified into one of three categories based on the Rothwell classification system: fact, policy, or value. The answers for each category were evaluated for reliability, quality and readability using The Journal of the American Medical Association Benchmark criteria, DISCERN score, Flesch-Kincaid Reading Ease Score and Grade Level. Results The Journal of the American Medical Association Benchmark criteria score for all three categories was 0, which is the lowest score indicating no reliable resources cited. The DISCERN score was 51 for fact, 53 for policy, and 55 for value questions, all of which are considered good scores. Across question categories, the reliability portion of the DISCERN score was low, due to a lack of resources. The Flesch-Kincaid Reading Ease Score (and Flesch-Kincaid Grade Level) was 48.3 (10.3) for the fact class, 42.0 (10.9) for the policy class, and 38.4 (11.6) for the value class. Conclusion The quality of information provided by the open AI chat system was generally high across all question types but had significant shortcomings in reliability due to the absence of source material citations. The DISCERN scores of the AI generated responses matched or exceeded previously published results of studies evaluating the quality of online information about rotator cuff repairs. The responses were U.S. 10th grade or higher reading level which is above the AMA and NIH recommendation of 6th grade reading level for patient materials. The AI software commonly referred the user to seek advice from orthopedic surgeons to improve their chances of a successful outcome.
Collapse
Affiliation(s)
- Eric Warren
- Duke University School of Medicine, Duke University, Durham, NC, USA
| | - Eoghan T. Hurley
- Department of Orthopaedic Surgery, Duke University, Durham, NC, USA
| | - Caroline N. Park
- Department of Orthopaedic Surgery, Duke University, Durham, NC, USA
| | - Bryan S. Crook
- Department of Orthopaedic Surgery, Duke University, Durham, NC, USA
| | - Samuel Lorentz
- Department of Orthopaedic Surgery, Duke University, Durham, NC, USA
| | - Jay M. Levin
- Department of Orthopaedic Surgery, Duke University, Durham, NC, USA
| | - Oke Anakwenze
- Department of Orthopaedic Surgery, Duke University, Durham, NC, USA
| | - Peter B. MacDonald
- Section of Orthopaedic Surgery & The Pan Am Clinic, University of Manitoba, Winnipeg, MB, Canada
| | | |
Collapse
|
108
|
Adhikari K, Naik N, Hameed BZ, Raghunath SK, Somani BK. Exploring the Ethical, Legal, and Social Implications of ChatGPT in Urology. Curr Urol Rep 2024; 25:1-8. [PMID: 37735339 DOI: 10.1007/s11934-023-01185-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/05/2023] [Indexed: 09/23/2023]
Abstract
PURPOSE OF THE REVIEW ChatGPT is programmed to generate responses based on pattern recognition. With this vast popularity and exponential growth, the question arises of moral issues, security and legitimacy. In this review article, we aim to analyze the ethical and legal implications of using ChatGPT in Urology and explore potential solutions addressing these concerns. RECENT FINDINGS There are many potential applications of ChatGPT in urology, and the extent to which it might improve healthcare may cause a profound shift in the way we deliver our services to patients and the overall healthcare system. This encompasses diagnosis and treatment planning, clinical workflow, patient education, augmenting consultations, and urological research. The ethical and legal considerations include patient autonomy and informed consent, privacy and confidentiality, bias and fairness, human oversight and accountability, trust and transparency, liability and malpractice, intellectual property rights, and regulatory framework. The application of ChatGPT in urology has shown great potential to improve patient care and assist urologists in various aspects of clinical practice, research, and education. Complying with data security and privacy regulations, and ensuring human oversight and accountability are some potential solutions to these legal and ethical concerns. Overall, the benefits and risks of using ChatGPT in urology must be weighed carefully, and a cautious approach must be taken to ensure that its use aligns with human values and advances patient care ethically and responsibly.
Collapse
Affiliation(s)
- Kinju Adhikari
- Department of Urology, HCG Cancer Centre, Bangaluru, India
| | - Nithesh Naik
- Department of Mechanical and Industrial Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, India
| | - Bm Zeeshan Hameed
- Department of Urology, Father Muller Medical College, Mangalore, Karnataka, India
| | - S K Raghunath
- Department of Urology, HCG Cancer Centre, Bangaluru, India
| | - Bhaskar K Somani
- Department of Urology, University Hospital Southampton NHS Trust, Southampton, SO16 6YD, UK.
| |
Collapse
|
109
|
Miao J, Thongprayoon C, Suppadungsuk S, Garcia Valencia OA, Qureshi F, Cheungpasitporn W. Ethical Dilemmas in Using AI for Academic Writing and an Example Framework for Peer Review in Nephrology Academia: A Narrative Review. Clin Pract 2023; 14:89-105. [PMID: 38248432 PMCID: PMC10801601 DOI: 10.3390/clinpract14010008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Revised: 12/23/2023] [Accepted: 12/28/2023] [Indexed: 01/23/2024] Open
Abstract
The emergence of artificial intelligence (AI) has greatly propelled progress across various sectors including the field of nephrology academia. However, this advancement has also given rise to ethical challenges, notably in scholarly writing. AI's capacity to automate labor-intensive tasks like literature reviews and data analysis has created opportunities for unethical practices, with scholars incorporating AI-generated text into their manuscripts, potentially undermining academic integrity. This situation gives rise to a range of ethical dilemmas that not only question the authenticity of contemporary academic endeavors but also challenge the credibility of the peer-review process and the integrity of editorial oversight. Instances of this misconduct are highlighted, spanning from lesser-known journals to reputable ones, and even infiltrating graduate theses and grant applications. This subtle AI intrusion hints at a systemic vulnerability within the academic publishing domain, exacerbated by the publish-or-perish mentality. The solutions aimed at mitigating the unethical employment of AI in academia include the adoption of sophisticated AI-driven plagiarism detection systems, a robust augmentation of the peer-review process with an "AI scrutiny" phase, comprehensive training for academics on ethical AI usage, and the promotion of a culture of transparency that acknowledges AI's role in research. This review underscores the pressing need for collaborative efforts among academic nephrology institutions to foster an environment of ethical AI application, thus preserving the esteemed academic integrity in the face of rapid technological advancements. It also makes a plea for rigorous research to assess the extent of AI's involvement in the academic literature, evaluate the effectiveness of AI-enhanced plagiarism detection tools, and understand the long-term consequences of AI utilization on academic integrity. An example framework has been proposed to outline a comprehensive approach to integrating AI into Nephrology academic writing and peer review. Using proactive initiatives and rigorous evaluations, a harmonious environment that harnesses AI's capabilities while upholding stringent academic standards can be envisioned.
Collapse
Affiliation(s)
- Jing Miao
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (S.S.); (O.A.G.V.); (F.Q.); (W.C.)
| | - Charat Thongprayoon
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (S.S.); (O.A.G.V.); (F.Q.); (W.C.)
| | - Supawadee Suppadungsuk
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (S.S.); (O.A.G.V.); (F.Q.); (W.C.)
- Chakri Naruebodindra Medical Institute, Faculty of Medicine Ramathibodi Hospital, Mahidol University, Bang Phli 10540, Samut Prakan, Thailand
| | - Oscar A. Garcia Valencia
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (S.S.); (O.A.G.V.); (F.Q.); (W.C.)
| | - Fawad Qureshi
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (S.S.); (O.A.G.V.); (F.Q.); (W.C.)
| | - Wisit Cheungpasitporn
- Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA; (J.M.); (S.S.); (O.A.G.V.); (F.Q.); (W.C.)
| |
Collapse
|
110
|
BaHammam AS. Balancing Innovation and Integrity: The Role of AI in Research and Scientific Writing. Nat Sci Sleep 2023; 15:1153-1156. [PMID: 38170140 PMCID: PMC10759812 DOI: 10.2147/nss.s455765] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Accepted: 12/21/2023] [Indexed: 01/05/2024] Open
Affiliation(s)
- Ahmed S BaHammam
- Editor-in-Chief Nature and Science of Sleep
- Department of Medicine, University Sleep Disorders Center and Pulmonary Service, King Saud University, Riyadh, Saudi Arabia
- King Saud University Medical City, Riyadh, Saudi Arabia
| |
Collapse
|
111
|
Kim TW. Application of artificial intelligence chatbots, including ChatGPT, in education, scholarly work, programming, and content generation and its prospects: a narrative review. JOURNAL OF EDUCATIONAL EVALUATION FOR HEALTH PROFESSIONS 2023; 20:38. [PMID: 38148495 DOI: 10.3352/jeehp.2023.20.38] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/26/2023] [Accepted: 12/26/2023] [Indexed: 12/28/2023]
Abstract
This study aims to explore ChatGPT’s (GPT-3.5 version) functionalities, including reinforcement learning, diverse applications, and limitations. ChatGPT is an artificial intelligence (AI) chatbot powered by OpenAI’s Generative Pre-trained Transformer (GPT) model. The chatbot’s applications span education, programming, content generation, and more, demonstrating its versatility. ChatGPT can improve education by creating assignments and offering personalized feedback, as shown by its notable performance in medical exams and the United States Medical Licensing Exam. However, concerns include plagiarism, reliability, and educational disparities. It aids in various research tasks, from design to writing, and has shown proficiency in summarizing and suggesting titles. Its use in scientific writing and language translation is promising, but professional oversight is needed for accuracy and originality. It assists in programming tasks like writing code, debugging, and guiding installation and updates. It offers diverse applications, from cheering up individuals to generating creative content like essays, news articles, and business plans. Unlike search engines, ChatGPT provides interactive, generative responses and understands context, making it more akin to human conversation, in contrast to conventional search engines’ keyword-based, non-interactive nature. ChatGPT has limitations, such as potential bias, dependence on outdated data, and revenue generation challenges. Nonetheless, ChatGPT is considered to be a transformative AI tool poised to redefine the future of generative technology. In conclusion, advancements in AI, such as ChatGPT, are altering how knowledge is acquired and applied, marking a shift from search engines to creativity engines. This transformation highlights the increasing importance of AI literacy and the ability to effectively utilize AI in various domains of life.
Collapse
Affiliation(s)
- Tae Won Kim
- AI‧Future Strategy Center, National Information Society Agency of Korea, Daegu, Korea
| |
Collapse
|
112
|
Liang J, Wang L, Luo J, Yan Y, Fan C. The relationship between student interaction with generative artificial intelligence and learning achievement: serial mediating roles of self-efficacy and cognitive engagement. Front Psychol 2023; 14:1285392. [PMID: 38187430 PMCID: PMC10766754 DOI: 10.3389/fpsyg.2023.1285392] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Accepted: 11/28/2023] [Indexed: 01/09/2024] Open
Abstract
Generative artificial intelligence (GAI) shocked the world with its unprecedented ability and raised significant tensions in the education field. Educators inevitably transition to an educational future that embraces GAI rather than shuns it. Understanding the mechanism between students interacting with GAI tools and their achievement is important for educators and schools, but relevant empirical evidence is relatively lacking. Due to the characteristics of personalization and real-time interactivity of GAI tools, we propose that the students-GAI interaction would affect their learning achievement through serial mediators of self-efficacy and cognitive engagement. Based on questionnaire surveys that include 389 participants as the objective, this study finds that: (1) in total, there is a significantly positive relationship between student-GAI interaction and learning achievement. (2) This positive relationship is mediated by self-efficacy, with a significant mediation effect value of 0.015. (3) Cognitive engagement also acts as a mediator in the mechanism between the student-GAI interaction and learning achievement, evidenced by a significant and relatively strong mediating effect value of 0.046. (4) Self-efficacy and cognitive engagement in series mediate this positive association, with a serial mediating effect value of 0.011, which is relatively small in comparison but also shows significance. In addition, the propensity score matching (PSM) method is applied to alleviate self-selection bias, reinforcing the validity of the results. The findings offer empirical evidence for the incorporation of GAI in teaching and learning.
Collapse
Affiliation(s)
- Jing Liang
- College of Management Science, Chengdu University of Technology, Chengdu, China
| | - Lili Wang
- School of Logistics, Chengdu University of Information Technology, Chengdu, China
| | - Jia Luo
- Business School, Chengdu University, Chengdu, China
| | - Yufei Yan
- Business School, Southwest Minzu University, Chengdu, China
| | - Chao Fan
- College of Management Science, Chengdu University of Technology, Chengdu, China
| |
Collapse
|
113
|
Anghelescu A, Ciobanu I, Munteanu C, Anghelescu LAM, Onose G. ChatGPT: "To be or not to be" ... in academic research. The human mind's analytical rigor and capacity to discriminate between AI bots' truths and hallucinations. BALNEO AND PRM RESEARCH JOURNAL 2023; 14:614. [DOI: 10.12680/balneo.2023.614] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/11/2025] Open
Abstract
Background. ChatGPT can generate increasingly realistic language, but the correctness and integrity of implementing these models in scientific papers remain unknown.
Recently published literature emphasized the ”three faces of the coin” of ChatGPT: the negative impact on academic writing, limitations in analyzing and conducting extensive searches of references across multiple databases, and the superiority of the human mind.
Method. The present study assessed the chatbot's ability for improvement and its propensity for self-correction at various points in 2023.
Starting from previous papers published in our clinic, the authors repeatedly challenged the ChatGPT to conduct extensive searches for references across multiple databases at different time intervals (in March and September 2023). The bot was asked to find recent meta-analyses on a particular topic.
Results. The replies (print screens) generated in March and September 2023 serve as evidence of the OpenAI platform's qualitative development and improvement.
During the first contact with ChatGPT-3, one noticed significant content flows and drawbacks. ChatGPT provided references and short essays, but none of them were real, despite ChatGPT's clear affirmative response. When searching PubMed IDs, all DOI numbers indicated by the chatbot correlated to various unconnected manuscripts.
After a few months, the authors repeated the same interrogative provocations and observed a significant shift in the replies. The ChatGPT-3.5 delivered balanced responses, emphasizing the superiority of the human intellect and advocating traditional academic research techniques and methods.
Discussion. A recent comparative systematic analysis using the PRISMA method using the same keyword syntactic correlations to search for systematic literature or open sources has revealed the superiority of the classical scholarly method of research.
In contrast, every document (title, authors, doi) that ChatGPT-3 initially delivered was erroneous and associated with a different field or topic.
Literature published during the first trimester of 2023 emphasized ChatGPT`s hallucinatory tendency to supply fake ”bibliographic resources” and confabulatory attempts to paraphrase nonexistent ”research papers” presented as authentic articles.
A second inquiry was realized six months later generated reserved and cautious solutions, indicating the researcher should analyze and carefully verify the information from specialized academic databases.
Conclusions. The paper succinctly describes the flows and initial limitations of the ChatGPT-3 version and the process of updating and improving the GPT-3.5 system during 2023.
ChatGPT might be a possible adjunct to academic writing and scientific research, considering any limitations that might jeopardize the study.
The new perspective from ChatGPT claims that human intelligence and thought must thoroughly assess any AI information.
Collapse
Affiliation(s)
- Aurelian Anghelescu
- “Carol Davila” University of Medicine and Pharmacy, Bucharest, Romania; Teaching Emergency Hospital “Bagdasar-Arseni” (TEHBA), Bucharest, Romania
| | - Ilinca Ciobanu
- Teaching Emergency Hospital “Bagdasar-Arseni” (TEHBA), Bucharest, Romania
| | | | | | - Gelu Onose
- “Carol Davila” University of Medicine and Pharmacy, Bucharest, Romania; Teaching Emergency Hospital “Bagdasar-Arseni” (TEHBA), Bucharest, Romania
| |
Collapse
|
114
|
Affiliation(s)
- Linda Harrington
- Linda Harrington is an Independent Consultant, Health Informatics and Digital Strategy, and Adjunct Faculty at Texas Christian University, 2800 South University Drive, Fort Worth, TX 76109
| |
Collapse
|
115
|
Tang G. Academic journals cannot simply require authors to declare that they used ChatGPT. Ir J Med Sci 2023; 192:3195-3196. [PMID: 37058235 DOI: 10.1007/s11845-023-03374-x] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2023] [Accepted: 04/11/2023] [Indexed: 04/15/2023]
Abstract
This letter to the editor points out weaknesses in the editorial policies of some academic journals regarding the use of ChatGPT-generated content. Editorial policies should provide more specific details on which parts of an academic paper are allowed to use ChatGPT-generated content. If authors use ChatGPT-generated content in the conclusion or results section, it may harm the academic paper's originality and, therefore, should not be accepted.
Collapse
Affiliation(s)
- Gengyan Tang
- Sichuan Academy of Social Sciences, Chengdu, China.
| |
Collapse
|
116
|
Wei Y, Guo X. Revolutionizing clinical experimental protocol design through the ChatGPT technology. Am J Med Sci 2023; 366:468-470. [PMID: 37699445 DOI: 10.1016/j.amjms.2023.09.004] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2023] [Revised: 09/01/2023] [Accepted: 09/05/2023] [Indexed: 09/14/2023]
Affiliation(s)
- Yanhui Wei
- School of Medicine, Southeast University, Nanjing, China
| | - Xuejun Guo
- Department of Haematology, Puyang Oilfield General Hospital Affiliated with Xinxiang Medical University, Puyang, China; Puyang Translational Medicine Engineering and Technology Research Center, Puyang, China.
| |
Collapse
|
117
|
Levin G, Brezinov Y, Meyer R. Exploring the use of ChatGPT in OBGYN: a bibliometric analysis of the first ChatGPT-related publications. Arch Gynecol Obstet 2023; 308:1785-1789. [PMID: 37222839 DOI: 10.1007/s00404-023-07081-x] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Accepted: 05/08/2023] [Indexed: 05/25/2023]
Abstract
PURPOSE Little is known about the scientific literature regarding the new revolutionary tool, ChatGPT. We aim to perform a bibliometric analysis to identify ChatGPT-related publications in obstetrics and gynecology (OBGYN). STUDY DESIGN A bibliometric study through PubMed database. We mined all ChatGPT-related publications using the search term "ChatGPT". Bibliometric data were obtained from the iCite database. We performed a descriptive analysis. We further compared IF among publications describing a study vs. other publications. RESULTS Overall, 42 ChatGPT-related publications were published across 26 different journals during 69 days. Most publications were editorials (52%) and news/briefing (22%), with only one (2%) research article identified. Five (12%) publications described a study performed. No ChatGPT-related publications in OBGYN were found. The leading journal by the number of publications was Nature (24%), followed by Lancet Digital Health and Radiology (7%, for both). The main subjects of publications were ChatGPT's scientific writing quality (26%) and a description of ChatGPT (26%) followed by tested performance of ChatGPT (14%), authorship and ethical issues (10% for both topics).In a comparison of publications describing a study performed (n = 5) vs. other publications (n = 37), mean IF was lower in the study-publications (mean 6.25 ± 0 vs. 25.4 ± 21.6, p < .001). CONCLUSIONS The study highlights main trends in ChatGPT-related publications. OBGYN is yet to be represented in this literature.
Collapse
Affiliation(s)
- Gabriel Levin
- The Department of Gynecologic Oncology, Hadassah-Hebrew University Medical Center, Jerusalem, Israel.
- Lady Davis Institute for Cancer Research, Jewish General Hospital, McGill University, Quebec, Canada.
| | - Yoav Brezinov
- Experimental Surgery, McGill University, Quebec, Canada
| | - Raanan Meyer
- Division of Minimally Invasive Gynecologic Surgery, Department of Obstetrics and Gynecology, Cedars Sinai Medical Center, Los Angeles, CA, USA
- The Dr. Pinchas Bornstein Talpiot Medical Leadership Program, Sheba Medical Center, Tel Hashomer, Ramat-Gan, Israel
| |
Collapse
|
118
|
Bisi T, Risser A, Clavert P, Migaud H, Dartus J. What is the rate of text generated by artificial intelligence over a year of publication in Orthopedics & Traumatology: Surgery & Research? Analysis of 425 articles before versus after the launch of ChatGPT in November 2022. Orthop Traumatol Surg Res 2023; 109:103694. [PMID: 37776949 DOI: 10.1016/j.otsr.2023.103694] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/17/2023] [Revised: 07/10/2023] [Accepted: 08/24/2023] [Indexed: 10/02/2023]
Abstract
BACKGROUND The use of artificial intelligence (AI) is soaring, and the launch of ChatGPT in November 2022 has accelerated this trend. This "chatbot" can generate complete scientific articles, with risk of plagiarism by mining existing data or downright fraud by fabricating studies with no real data at all. There are tools that detect AI in publications, but to our knowledge they have not been systematically assessed for publication in scientific journals. We therefore conducted a retrospective study on articles published in Orthopaedics & Traumatology: Surgery & Research (OTSR): firstly, to screen for AI-generated content before and after the publicized launch of ChatGPT; secondly, to assess whether AI was more often used in some countries than others to generate content; thirdly, to determine whether plagiarism rate correlated with AI-generation, and lastly, to determine whether elements other than text generation, and notably the translation procedure, could raise suspicion of AI use. HYPOTHESIS The rate of AI use increased after the publicized launch of ChatGPT v3.5 in November 2022. MATERIAL AND METHODS In all, 425 articles published between February 2022 and September 2023 (221 before and 204 after November 1, 2022) underwent ZeroGPT assessment of the level of AI generation in the final English-language version (abstract and body of the article). Two scores were obtained: probability of AI generation, in six grades from Human to AI; and percentage AI generation. Plagiarism was assessed on the Ithenticate application at submission. Articles in French were assessed in their English-language version as translated by a human translator, with comparison to automatic translation by Google Translate and DeepL. RESULTS AI-generated text was detected mainly in Abstracts, with a 10.1% rate of AI or considerable AI generation, compared to only 1.9% for the body of the article and 5.6% for the total body+abstract. Analysis for before and after November 2022 found an increase in AI generation in body+abstract, from 10.30±15.95% (range, 0-100%) to 15.64±19.8% (range, 0-99.93) (p < 0.04; NS for abstracts alone). AI scores differed between types of article: 14.9% for original articles and 9.8% for reviews (p<0.01). The highest rates of probable AI generation were in articles from Japan, China, South America and English-speaking countries (p<0.0001). Plagiarism rates did not increase between the two study periods, and were unrelated to AI rates. On the other hand, when articles were classified as "suspected" of AI generation (plagiarism rate ≥ 20%) or "non-suspected" (rate<20%), the "similarity" score was higher in suspect articles: 25.7±13.23% (range, 10-69%) versus 16.28±10% (range, 0-79%) (p < 0.001). In the body of the article, use of translation software was associated with higher AI rates than with a human translator: 3.5±5% for human translators, versus 18±10% and 21.9±11% respectively for Google Translate and DeepL (p < 0.001). DISCUSSION The present study revealed an increasing rate of AI use in articles published in OTSR. AI grades differed according to type of article and country of origin. Use of translation software increased the AI grade. In the long run, use of ChatGPT incurs a risk of plagiarism and scientific misconduct, and needs to be detected and signaled by a digital tag on any robot-generated text. LEVEL OF EVIDENCE III; case-control study.
Collapse
Affiliation(s)
- Théophile Bisi
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France.
| | - Anthony Risser
- Service de chirurgie du membre supérieur, Hautepierre 2, CHRU Strasbourg, 1, avenue Molière, 67200 Strasbourg, France
| | - Philippe Clavert
- Service de chirurgie du membre supérieur, Hautepierre 2, CHRU Strasbourg, 1, avenue Molière, 67200 Strasbourg, France; Faculté de médecine, institut d'anatomie normale, 4, rue Kirschleger, 67085 Strasbourg, France
| | - Henri Migaud
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France
| | - Julien Dartus
- Département universitaire de chirurgie orthopédique, université de Lille, CHU de Lille, 59000 Lille, France; Service de chirurgie orthopédique, centre hospitalier universitaire (CHU) de Lille, hôpital Roger-Salengro, place de Verdun, 59000 Lille, France
| |
Collapse
|
119
|
Chatterjee S, Bhattacharya M, Pal S, Lee SS, Chakraborty C. ChatGPT and large language models in orthopedics: from education and surgery to research. J Exp Orthop 2023; 10:128. [PMID: 38038796 PMCID: PMC10692045 DOI: 10.1186/s40634-023-00700-1] [Citation(s) in RCA: 25] [Impact Index Per Article: 12.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Accepted: 11/16/2023] [Indexed: 12/02/2023] Open
Abstract
ChatGPT has quickly popularized since its release in November 2022. Currently, large language models (LLMs) and ChatGPT have been applied in various domains of medical science, including in cardiology, nephrology, orthopedics, ophthalmology, gastroenterology, and radiology. Researchers are exploring the potential of LLMs and ChatGPT for clinicians and surgeons in every domain. This study discusses how ChatGPT can help orthopedic clinicians and surgeons perform various medical tasks. LLMs and ChatGPT can help the patient community by providing suggestions and diagnostic guidelines. In this study, the use of LLMs and ChatGPT to enhance and expand the field of orthopedics, including orthopedic education, surgery, and research, is explored. Present LLMs have several shortcomings, which are discussed herein. However, next-generation and future domain-specific LLMs are expected to be more potent and transform patients' quality of life.
Collapse
Affiliation(s)
- Srijan Chatterjee
- Institute for Skeletal Aging & Orthopaedic Surgery, Hallym University-Chuncheon Sacred Heart Hospital, Chuncheon-Si, 24252, Gangwon-Do, Republic of Korea
| | - Manojit Bhattacharya
- Department of Zoology, Fakir Mohan University, Vyasa Vihar, Balasore, 756020, Odisha, India
| | - Soumen Pal
- School of Mechanical Engineering, Vellore Institute of Technology, Vellore, Tamil Nadu, India
| | - Sang-Soo Lee
- Institute for Skeletal Aging & Orthopaedic Surgery, Hallym University-Chuncheon Sacred Heart Hospital, Chuncheon-Si, 24252, Gangwon-Do, Republic of Korea.
| | - Chiranjib Chakraborty
- Department of Biotechnology, School of Life Science and Biotechnology, Adamas University, Kolkata, West Bengal, 700126, India.
| |
Collapse
|
120
|
Rossettini G, Cook C, Palese A, Pillastrini P, Turolla A. Pros and Cons of Using Artificial Intelligence Chatbots for Musculoskeletal Rehabilitation Management. J Orthop Sports Phys Ther 2023; 53:728-734. [PMID: 37707390 DOI: 10.2519/jospt.2023.12000] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 09/15/2023]
Abstract
SYNOPSIS: Artificial intelligence (AI), specifically large language models (LLMs), which focus on the interaction between computers and human language, can influence musculoskeletal rehabilitation management. AI chatbots (eg, ChatGPT, Microsoft Bing, and Google Bard) are a form of large language models designed to understand, interpret, and generate text similar to what is produced by humans. Since their release, chatbots have triggered controversy in the international scientific community, including when they have passed university exams, generated credible scientific abstracts, and shown potential for replacing humans in scientific roles. The controversies extend to the field of musculoskeletal rehabilitation. In this Viewpoint, we describe the potential applications and limitations, and recommended actions for education, clinical practice, and research when using AI chatbots for musculoskeletal rehabilitation management, aspects that may have similar implications for the broader health care community. J Orthop Sports Phys Ther 2023;53(12):1-7. Epub 14 September 2023. doi:10.2519/jospt.2023.12000.
Collapse
|
121
|
Tate S, Fouladvand S, Chen JH, Chen CYA. The ChatGPT therapist will see you now: Navigating generative artificial intelligence's potential in addiction medicine research and patient care. Addiction 2023; 118:2249-2251. [PMID: 37735091 DOI: 10.1111/add.16341] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Accepted: 08/23/2023] [Indexed: 09/23/2023]
Affiliation(s)
- Steven Tate
- Department of Psychiatry and Behavioural Sciences, Stanford University School of Medicine, Palo Alto, California, USA
| | - Sajjad Fouladvand
- Department of Medicine, Stanford University School of Medicine, Palo Alto, California, USA
| | - Jonathan H Chen
- Department of Medicine, Stanford University School of Medicine, Palo Alto, California, USA
| | - Chwen-Yuen Angie Chen
- Department of Medicine, Stanford University School of Medicine, Palo Alto, California, USA
| |
Collapse
|
122
|
Kurian N, James D, Varghese VS, Cherian JM, Varghese KG. Artificial intelligence in scientific publications. J Am Dent Assoc 2023; 154:1041-1043. [PMID: 37140497 DOI: 10.1016/j.adaj.2023.03.015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Revised: 03/16/2023] [Accepted: 03/27/2023] [Indexed: 05/05/2023]
|
123
|
Abstract
Since its launch, ChatGPT, an artificial intelligence-powered language model tool, has generated significant attention in research writing. The use of ChatGPT in medical research can be a double-edged sword. ChatGPT can expedite the research writing process by assisting with hypothesis formulation, literature review, data analysis and manuscript writing. On the other hand, using ChatGPT raises concerns regarding the originality and authenticity of content, the precision and potential bias of the tool's output, and the potential legal issues associated with privacy, confidentiality and plagiarism. The article also calls for adherence to stringent citation guidelines and the development of regulations promoting the responsible application of AI. Despite the revolutionary capabilities of ChatGPT, the article highlights its inability to replicate human thought and the difficulties in maintaining the integrity and reliability of ChatGPT-enabled research, particularly in complex fields such as medicine and law. AI tools can be used as supplementary aids rather than primary sources of analysis in medical research writing.
Collapse
Affiliation(s)
- Daideepya C Bhargava
- Department of Forensic Medicine and Toxicology, All India Institute of Medical Sciences, India
| | - Devendra Jadav
- Department of Forensic Medicine and Toxicology, All India Institute of Medical Sciences, India
| | - Vikas P Meshram
- Department of Forensic Medicine and Toxicology, All India Institute of Medical Sciences, India
| | - Tanuj Kanchan
- Department of Forensic Medicine and Toxicology, All India Institute of Medical Sciences, India
| |
Collapse
|
124
|
Marra AR, Nori P, Langford BJ, Kobayashi T, Bearman G. Brave new world: Leveraging artificial intelligence for advancing healthcare epidemiology, infection prevention, and antimicrobial stewardship. Infect Control Hosp Epidemiol 2023; 44:1909-1912. [PMID: 37395009 DOI: 10.1017/ice.2023.122] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/04/2023]
Affiliation(s)
- Alexandre R Marra
- Hospital Israelita Albert Einstein, São Paulo, Brazil
- Department of Internal Medicine, University of Iowa Carver College of Medicine, Iowa City, Iowa, United States
| | - Priya Nori
- Division of Infectious Diseases, Department of Medicine, Montefiore Health System, Albert Einstein College of Medicine, Bronx, New York, United States
| | - Bradley J Langford
- Dalla Lana School of Public Health, University of Toronto, Toronto, Canada
- Hotel Dieu Shaver Health and Rehabilitation Centre, St. Catharines, Canada
| | - Takaaki Kobayashi
- Department of Internal Medicine, University of Iowa Carver College of Medicine, Iowa City, Iowa, United States
| | - Gonzalo Bearman
- Division of Infectious Diseases, Virginia Commonwealth University Health, Virginia Commonwealth University, Richmond, Virginia, United States
| |
Collapse
|
125
|
Silva TP, Ocampo TSC, Alencar-Palha C, Oliveira-Santos C, Takeshita WM, Oliveira ML. ChatGPT: a tool for scientific writing or a threat to integrity? Br J Radiol 2023; 96:20230430. [PMID: 37750843 PMCID: PMC10646664 DOI: 10.1259/bjr.20230430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2023] [Revised: 08/03/2023] [Accepted: 08/06/2023] [Indexed: 09/27/2023] Open
Abstract
The use of ChatGPT as a tool for writing and knowledge integration raises concerns about the potential for its use to replace critical thinking and academic writing skills. While ChatGPT can assist in generating text and suggesting appropriate language, it should not replace the human responsibility for creating innovative knowledge through experiential learning. The accuracy and quality of information provided by ChatGPT also require caution, as previous studies have reported inaccuracies in references used by chatbots. ChatGPT acknowledges certain limitations, including the potential for generating erroneous or biased content, and it is essential to exercise caution in interpreting its responses and recognize the indispensable role of human experience in the processes of information retrieval and knowledge creation. Furthermore, the challenge of distinguishing between papers written by humans or AI highlights the need for thorough review processes to prevent the spread of articles that could lead to the loss of confidence in the accuracy and integrity of scientific research. Overall, while the use of ChatGPT can be helpful, it is crucial to raise awareness of the potential issues associated with the use of ChatGPT, as well as to discuss boundaries so that AI can be used without compromising the quality of scientific articles and the integrity of evidence-based knowledge.
Collapse
Affiliation(s)
- Thaísa Pinheiro Silva
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| | - Thaís S C Ocampo
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| | - Caio Alencar-Palha
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| | - Christiano Oliveira-Santos
- Department of Diagnosis and Oral Health, University of Louisville School of Dentistry, Louisville, United States
| | - Wilton Mitsunari Takeshita
- Department of Diagnosis and Surgery, Paulista State University Júlio de Mesquita Filho, Araçatuba, Brazil
| | - Matheus L Oliveira
- Department of Oral Diagnosis, Division of Oral Radiology, Piracicaba Dental School, University of Campinas, Piracicaba, Brazil
| |
Collapse
|
126
|
Berenbaum MR. Welcome, AI overlords? Proc Natl Acad Sci U S A 2023; 120:e2318980120. [PMID: 37991944 PMCID: PMC10691220 DOI: 10.1073/pnas.2318980120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2023] Open
|
127
|
Tian S, Jin Q, Yeganova L, Lai PT, Zhu Q, Chen X, Yang Y, Chen Q, Kim W, Comeau DC, Islamaj R, Kapoor A, Gao X, Lu Z. Opportunities and challenges for ChatGPT and large language models in biomedicine and health. Brief Bioinform 2023; 25:bbad493. [PMID: 38168838 PMCID: PMC10762511 DOI: 10.1093/bib/bbad493] [Citation(s) in RCA: 83] [Impact Index Per Article: 41.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2023] [Revised: 11/15/2023] [Accepted: 12/06/2023] [Indexed: 01/05/2024] Open
Abstract
ChatGPT has drawn considerable attention from both the general public and domain experts with its remarkable text generation capabilities. This has subsequently led to the emergence of diverse applications in the field of biomedicine and health. In this work, we examine the diverse applications of large language models (LLMs), such as ChatGPT, in biomedicine and health. Specifically, we explore the areas of biomedical information retrieval, question answering, medical text summarization, information extraction and medical education and investigate whether LLMs possess the transformative power to revolutionize these tasks or whether the distinct complexities of biomedical domain presents unique challenges. Following an extensive literature survey, we find that significant advances have been made in the field of text generation tasks, surpassing the previous state-of-the-art methods. For other applications, the advances have been modest. Overall, LLMs have not yet revolutionized biomedicine, but recent rapid progress indicates that such methods hold great potential to provide valuable means for accelerating discovery and improving health. We also find that the use of LLMs, like ChatGPT, in the fields of biomedicine and health entails various risks and challenges, including fabricated information in its generated responses, as well as legal and privacy concerns associated with sensitive patient data. We believe this survey can provide a comprehensive and timely overview to biomedical researchers and healthcare practitioners on the opportunities and challenges associated with using ChatGPT and other LLMs for transforming biomedicine and health.
Collapse
Affiliation(s)
- Shubo Tian
- National Library of Medicine, National Institutes of Health
| | - Qiao Jin
- National Library of Medicine, National Institutes of Health
| | - Lana Yeganova
- National Library of Medicine, National Institutes of Health
| | - Po-Ting Lai
- National Library of Medicine, National Institutes of Health
| | - Qingqing Zhu
- National Library of Medicine, National Institutes of Health
| | - Xiuying Chen
- King Abdullah University of Science and Technology
| | - Yifan Yang
- National Library of Medicine, National Institutes of Health
| | - Qingyu Chen
- National Library of Medicine, National Institutes of Health
| | - Won Kim
- National Library of Medicine, National Institutes of Health
| | | | | | - Aadit Kapoor
- National Library of Medicine, National Institutes of Health
| | - Xin Gao
- King Abdullah University of Science and Technology
| | - Zhiyong Lu
- National Library of Medicine, National Institutes of Health
| |
Collapse
|
128
|
Wong RSY, Ming LC, Raja Ali RA. The Intersection of ChatGPT, Clinical Medicine, and Medical Education. JMIR MEDICAL EDUCATION 2023; 9:e47274. [PMID: 37988149 DOI: 10.2196/47274] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/14/2023] [Revised: 06/16/2023] [Accepted: 06/30/2023] [Indexed: 11/22/2023]
Abstract
As we progress deeper into the digital age, the robust development and application of advanced artificial intelligence (AI) technology, specifically generative language models like ChatGPT (OpenAI), have potential implications in all sectors including medicine. This viewpoint article aims to present the authors' perspective on the integration of AI models such as ChatGPT in clinical medicine and medical education. The unprecedented capacity of ChatGPT to generate human-like responses, refined through Reinforcement Learning with Human Feedback, could significantly reshape the pedagogical methodologies within medical education. Through a comprehensive review and the authors' personal experiences, this viewpoint article elucidates the pros, cons, and ethical considerations of using ChatGPT within clinical medicine and notably, its implications for medical education. This exploration is crucial in a transformative era where AI could potentially augment human capability in the process of knowledge creation and dissemination, potentially revolutionizing medical education and clinical practice. The importance of maintaining academic integrity and professional standards is highlighted. The relevance of establishing clear guidelines for the responsible and ethical use of AI technologies in clinical medicine and medical education is also emphasized.
Collapse
Affiliation(s)
- Rebecca Shin-Yee Wong
- Department of Medical Education, School of Medical and Life Sciences, Sunway University, Selangor, Malaysia
- Faculty of Medicine, Nursing and Health Sciences, SEGi University, Petaling Jaya, Malaysia
| | - Long Chiau Ming
- School of Medical and Life Sciences, Sunway University, Selangor, Malaysia
| | - Raja Affendi Raja Ali
- School of Medical and Life Sciences, Sunway University, Selangor, Malaysia
- GUT Research Group, Faculty of Medicine, Universiti Kebangsaan Malaysia, Kuala Lumpur, Malaysia
| |
Collapse
|
129
|
Abuyaman O. Strengths and Weaknesses of ChatGPT Models for Scientific Writing About Medical Vitamin B12: Mixed Methods Study. JMIR Form Res 2023; 7:e49459. [PMID: 37948100 PMCID: PMC10674142 DOI: 10.2196/49459] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2023] [Revised: 08/17/2023] [Accepted: 10/29/2023] [Indexed: 11/12/2023] Open
Abstract
BACKGROUND ChatGPT is a large language model developed by OpenAI designed to generate human-like responses to prompts. OBJECTIVE This study aims to evaluate the ability of GPT-4 to generate scientific content and assist in scientific writing using medical vitamin B12 as the topic. Furthermore, the study will compare the performance of GPT-4 to its predecessor, GPT-3.5. METHODS The study examined responses from GPT-4 and GPT-3.5 to vitamin B12-related prompts, focusing on their quality and characteristics and comparing them to established scientific literature. RESULTS The results indicated that GPT-4 can potentially streamline scientific writing through its ability to edit language and write abstracts, keywords, and abbreviation lists. However, significant limitations of ChatGPT were revealed, including its inability to identify and address bias, inability to include recent information, lack of transparency, and inclusion of inaccurate information. Additionally, it cannot check for plagiarism or provide proper references. The accuracy of GPT-4's answers was found to be superior to GPT-3.5. CONCLUSIONS ChatGPT can be considered a helpful assistant in the writing process but not a replacement for a scientist's expertise. Researchers must remain aware of its limitations and use it appropriately. The improvements in consecutive ChatGPT versions suggest the possibility of overcoming some present limitations in the near future.
Collapse
Affiliation(s)
- Omar Abuyaman
- Department of Medical Laboratory Sciences, Faculty of Applied Medical Sciences, The Hashemite University, Zarqa, 13133, Jordan
| |
Collapse
|
130
|
Chlorogiannis DD, Apostolos A, Chlorogiannis A, Palaiodimos L, Giannakoulas G, Pargaonkar S, Xesfingi S, Kokkinidis DG. The Role of ChatGPT in the Advancement of Diagnosis, Management, and Prognosis of Cardiovascular and Cerebrovascular Disease. Healthcare (Basel) 2023; 11:2906. [PMID: 37958050 PMCID: PMC10648908 DOI: 10.3390/healthcare11212906] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2023] [Revised: 10/24/2023] [Accepted: 11/04/2023] [Indexed: 11/15/2023] Open
Abstract
Cardiovascular and cerebrovascular disease incidence has risen mainly due to poor control of preventable risk factors and still constitutes a significant financial and health burden worldwide. ChatGPT is an artificial intelligence language-based model developed by OpenAI. Due to the model's unique cognitive capabilities beyond data processing and the production of high-quality text, there has been a surge of research interest concerning its role in the scientific community and contemporary clinical practice. To fully exploit ChatGPT's potential benefits and reduce its possible misuse, extreme caution must be taken to ensure its implications ethically and equitably. In this narrative review, we explore the language model's possible applications and limitations while emphasizing its potential value for diagnosing, managing, and prognosis of cardiovascular and cerebrovascular disease.
Collapse
Affiliation(s)
| | - Anastasios Apostolos
- First Department of Cardiology, School of Medicine, National Kapodistrian University of Athens, Hippokrateion General Hospital of Athens, 115 27 Athens, Greece;
| | - Anargyros Chlorogiannis
- Department of Health Economics, Policy and Management, Karolinska Institutet, 171 77 Stockholm, Sweden
| | - Leonidas Palaiodimos
- Division of Hospital Medicine, Jacobi Medical Center, NYC H+H, Albert Einstein College of Medicine, New York, NY 10461, USA; (L.P.); (S.P.)
| | - George Giannakoulas
- Department of Cardiology, AHEPA University Hospital, Aristotle University of Thessaloniki, 541 24 Thessaloniki, Greece;
| | - Sumant Pargaonkar
- Division of Hospital Medicine, Jacobi Medical Center, NYC H+H, Albert Einstein College of Medicine, New York, NY 10461, USA; (L.P.); (S.P.)
| | - Sofia Xesfingi
- Department of Economics, University of Piraeus, 185 34 Piraeus, Greece
| | - Damianos G. Kokkinidis
- Section of Cardiovascular Medicine, Yale University School of Medicine, New Haven, CT 06510, USA
| |
Collapse
|
131
|
Peters V, Baumgartner M, Froese S, Morava E, Patterson M, Zschocke J, Rahman S. Risk and potential of ChatGPT in scientific publishing. J Inherit Metab Dis 2023; 46:1005-1006. [PMID: 37534774 DOI: 10.1002/jimd.12666] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 08/04/2023]
Affiliation(s)
- Verena Peters
- Centre for Paediatric and Adolescent Medicine, Metabolic Center, University Hospital Heidelberg, Heidelberg, Germany
| | - Matthias Baumgartner
- Division of Metabolism and Children's Research Center, University Children's Hospital Zurich, University of Zurich, Zurich, Switzerland
| | - Sean Froese
- Division of Metabolism and Children's Research Center, University Children's Hospital Zurich, University of Zurich, Zurich, Switzerland
| | - Eva Morava
- Department of Clinical Genomics, Mayo Clinic, Rochester, Minnesota, USA
- Department of Laboratory Medicine and Pathology, Mayo Clinic, Rochester, Minnesota, USA
| | - Marc Patterson
- Division of Child and Adolescent Neurology, Departments of Neurology, Pediatrics, and Clinical Genomics, Mayo Clinic, Rochester, Minnesota, USA
| | - Johannes Zschocke
- Institute of Human Genetics, Medical University Innsbruck, Innsbruck, Austria
| | - Shamima Rahman
- Mitochondrial Research Group, UCL Great Ormond Street Institute of Child Health, London, UK
- Metabolic Unit, Great Ormond Street Hospital for Children NHS Foundation Trust, London, UK
| |
Collapse
|
132
|
Braet F, Poger D. Let's have a chat about chatbot(s) in (biological) microscopy. J Microsc 2023; 292:59-63. [PMID: 37742291 DOI: 10.1111/jmi.13230] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2023] [Revised: 08/30/2023] [Accepted: 09/20/2023] [Indexed: 09/26/2023]
Affiliation(s)
- Filip Braet
- School of Medical Sciences (Molecular and Cellular Biomedicine), The University of Sydney, New South Wales, Australia
- Australian Centre for Microscopy and Microanalysis, The University of Sydney, Sydney, New South Wales, Australia
| | - David Poger
- Microscopy Australia, The University of Sydney, Sydney, New South Wales, Australia
| |
Collapse
|
133
|
Kaarre J, Feldt R, Keeling LE, Dadoo S, Zsidai B, Hughes JD, Samuelsson K, Musahl V. Exploring the potential of ChatGPT as a supplementary tool for providing orthopaedic information. Knee Surg Sports Traumatol Arthrosc 2023; 31:5190-5198. [PMID: 37553552 PMCID: PMC10598178 DOI: 10.1007/s00167-023-07529-2] [Citation(s) in RCA: 45] [Impact Index Per Article: 22.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Accepted: 07/26/2023] [Indexed: 08/10/2023]
Abstract
PURPOSE To investigate the potential use of large language models (LLMs) in orthopaedics by presenting queries pertinent to anterior cruciate ligament (ACL) surgery to generative pre-trained transformer (ChatGPT, specifically using its GPT-4 model of March 14th 2023). Additionally, this study aimed to evaluate the depth of the LLM's knowledge and investigate its adaptability to different user groups. It was hypothesized that the ChatGPT would be able to adapt to different target groups due to its strong language understanding and processing capabilities. METHODS ChatGPT was presented with 20 questions and response was requested for two distinct target audiences: patients and non-orthopaedic medical doctors. Two board-certified orthopaedic sports medicine surgeons and two expert orthopaedic sports medicine surgeons independently evaluated the responses generated by ChatGPT. Mean correctness, completeness, and adaptability to the target audiences (patients and non-orthopaedic medical doctors) were determined. A three-point response scale facilitated nuanced assessment. RESULTS ChatGPT exhibited fair accuracy, with average correctness scores of 1.69 and 1.66 (on a scale from 0, incorrect, 1, partially correct, to 2, correct) for patients and medical doctors, respectively. Three of the 20 questions (15.0%) were deemed incorrect by any of the four orthopaedic sports medicine surgeon assessors. Moreover, overall completeness was calculated to be 1.51 and 1.64 for patients and medical doctors, respectively, while overall adaptiveness was determined to be 1.75 and 1.73 for patients and doctors, respectively. CONCLUSION Overall, ChatGPT was successful in generating correct responses in approximately 65% of the cases related to ACL surgery. The findings of this study imply that LLMs offer potential as a supplementary tool for acquiring orthopaedic knowledge. However, although ChatGPT can provide guidance and effectively adapt to diverse target audiences, it cannot supplant the expertise of orthopaedic sports medicine surgeons in diagnostic and treatment planning endeavours due to its limited understanding of orthopaedic domains and its potential for erroneous responses. LEVEL OF EVIDENCE V.
Collapse
Affiliation(s)
- Janina Kaarre
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Göteborgsvägen 31, 431 80 Mölndal, Sweden
| | - Robert Feldt
- Department of Computer Science and Engineering, Chalmers University of Technology, Gothenburg, Sweden
| | - Laura E. Keeling
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| | - Sahil Dadoo
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| | - Bálint Zsidai
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Göteborgsvägen 31, 431 80 Mölndal, Sweden
| | - Jonathan D. Hughes
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| | - Kristian Samuelsson
- Department of Orthopaedics, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Göteborgsvägen 31, 431 80 Mölndal, Sweden
- Department of Orthopaedics, Sahlgrenska University Hospital, Mölndal, Sweden
| | - Volker Musahl
- Department of Orthopaedic Surgery, UPMC Freddie Fu Sports Medicine Center, University of Pittsburgh, Pittsburgh, USA
| |
Collapse
|
134
|
Torreggiani M, Piccoli GB, Mallett A. From the internet to the COVID-19 pandemic: how technological advances and a tumultuous world have changed scientific publishing and meetings. J Nephrol 2023; 36:2165-2167. [PMID: 37498469 DOI: 10.1007/s40620-023-01740-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/28/2023]
Affiliation(s)
- Massimo Torreggiani
- Néphrologie et dialyse, Centre Hospitalier Le Mans, 194 Avenue Rubillard, Le Mans, 72037, France.
| | - Giorgina Barbara Piccoli
- Néphrologie et dialyse, Centre Hospitalier Le Mans, 194 Avenue Rubillard, Le Mans, 72037, France
| | - Andrew Mallett
- Department of Renal Medicine, Townsville Hospital and Health Service, Townsville University Hospital, 100 Angus Smith Drive, Douglas, QLD, 4814, Australia
- College of Medicine and Dentistry, James Cook University, Townsville, QLD, Australia
- Institute for Molecular Bioscience, The University of Queensland, St Lucia, QLD, Australia
| |
Collapse
|
135
|
Preiksaitis C, Rose C. Opportunities, Challenges, and Future Directions of Generative Artificial Intelligence in Medical Education: Scoping Review. JMIR MEDICAL EDUCATION 2023; 9:e48785. [PMID: 37862079 PMCID: PMC10625095 DOI: 10.2196/48785] [Citation(s) in RCA: 60] [Impact Index Per Article: 30.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/07/2023] [Revised: 07/28/2023] [Accepted: 09/28/2023] [Indexed: 10/21/2023]
Abstract
BACKGROUND Generative artificial intelligence (AI) technologies are increasingly being utilized across various fields, with considerable interest and concern regarding their potential application in medical education. These technologies, such as Chat GPT and Bard, can generate new content and have a wide range of possible applications. OBJECTIVE This study aimed to synthesize the potential opportunities and limitations of generative AI in medical education. It sought to identify prevalent themes within recent literature regarding potential applications and challenges of generative AI in medical education and use these to guide future areas for exploration. METHODS We conducted a scoping review, following the framework by Arksey and O'Malley, of English language articles published from 2022 onward that discussed generative AI in the context of medical education. A literature search was performed using PubMed, Web of Science, and Google Scholar databases. We screened articles for inclusion, extracted data from relevant studies, and completed a quantitative and qualitative synthesis of the data. RESULTS Thematic analysis revealed diverse potential applications for generative AI in medical education, including self-directed learning, simulation scenarios, and writing assistance. However, the literature also highlighted significant challenges, such as issues with academic integrity, data accuracy, and potential detriments to learning. Based on these themes and the current state of the literature, we propose the following 3 key areas for investigation: developing learners' skills to evaluate AI critically, rethinking assessment methodology, and studying human-AI interactions. CONCLUSIONS The integration of generative AI in medical education presents exciting opportunities, alongside considerable challenges. There is a need to develop new skills and competencies related to AI as well as thoughtful, nuanced approaches to examine the growing use of generative AI in medical education.
Collapse
Affiliation(s)
- Carl Preiksaitis
- Department of Emergency Medicine, Stanford University School of Medicine, Palo Alto, CA, United States
| | - Christian Rose
- Department of Emergency Medicine, Stanford University School of Medicine, Palo Alto, CA, United States
| |
Collapse
|
136
|
Moons P, Van Bulck L. ChatGPT: can artificial intelligence language models be of value for cardiovascular nurses and allied health professionals. Eur J Cardiovasc Nurs 2023; 22:e55-e59. [PMID: 36752788 DOI: 10.1093/eurjcn/zvad022] [Citation(s) in RCA: 50] [Impact Index Per Article: 25.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Revised: 02/01/2022] [Accepted: 02/02/2023] [Indexed: 02/09/2023]
Affiliation(s)
- Philip Moons
- KU Leuven Department of Public Health and Primary Care, KU Leuven-University of Leuven, Kapucijnenvoer 35, Box 7001, 3000 Leuven, Belgium
- Institute of Health and Care Sciences, University of Gothenburg, Arvid Wallgrens backe 1, 413 46 Gothenburg, Sweden
- Department of Paediatrics and Child Health, University of Cape Town, Klipfontein Rd, Rondebosch, 7700 Cape Town, South Africa
| | - Liesbet Van Bulck
- KU Leuven Department of Public Health and Primary Care, KU Leuven-University of Leuven, Kapucijnenvoer 35, Box 7001, 3000 Leuven, Belgium
- Research Foundation Flanders (FWO), Leuvenseweg 38, 1000 Brussels, Belgium
| |
Collapse
|
137
|
Rashidi HH, Fennell BD, Albahra S, Hu B, Gorbett T. The ChatGPT conundrum: Human-generated scientific manuscripts misidentified as AI creations by AI text detection tool. J Pathol Inform 2023; 14:100342. [PMID: 38116171 PMCID: PMC10727991 DOI: 10.1016/j.jpi.2023.100342] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Revised: 10/08/2023] [Accepted: 10/10/2023] [Indexed: 12/21/2023] Open
Abstract
AI Chat Bots such as ChatGPT are revolutionizing our AI capabilities, especially in text generation, to help expedite many tasks, but they introduce new dilemmas. The detection of AI-generated text has become a subject of great debate considering the AI text detector's known and unexpected limitations. Thus far, much research in this area has focused on the detection of AI-generated text; however, the goal of this study was to evaluate the opposite scenario, an AI-text detection tool's ability to discriminate human-generated text. Thousands of abstracts from several of the most well-known scientific journals were used to test the predictive capabilities of these detection tools, assessing abstracts from 1980 to 2023. We found that the AI text detector erroneously identified up to 8% of the known real abstracts as AI-generated text. This further highlights the current limitations of such detection tools and argues for novel detectors or combined approaches that can address this shortcoming and minimize its unanticipated consequences as we navigate this new AI landscape.
Collapse
Affiliation(s)
- Hooman H. Rashidi
- Pathology and Laboratory Medicine Institute (PLMI), Cleveland Clinic, Cleveland, OH, United States
- PLMI’s Center for Artificial Intelligence & Data Science, Cleveland Clinic, Cleveland, OH, United States
| | - Brandon D. Fennell
- University of California, San Francisco – Department of Medicine, San Francisco, CA, United States
| | - Samer Albahra
- Pathology and Laboratory Medicine Institute (PLMI), Cleveland Clinic, Cleveland, OH, United States
- PLMI’s Center for Artificial Intelligence & Data Science, Cleveland Clinic, Cleveland, OH, United States
| | - Bo Hu
- Department of Quantitative Health Sciences, Cleveland Clinic, Cleveland, OH, United States
- PLMI’s Center for Artificial Intelligence & Data Science, Cleveland Clinic, Cleveland, OH, United States
| | - Tom Gorbett
- Pathology and Laboratory Medicine Institute (PLMI), Cleveland Clinic, Cleveland, OH, United States
- PLMI’s Center for Artificial Intelligence & Data Science, Cleveland Clinic, Cleveland, OH, United States
| |
Collapse
|
138
|
Clusmann J, Kolbinger FR, Muti HS, Carrero ZI, Eckardt JN, Laleh NG, Löffler CML, Schwarzkopf SC, Unger M, Veldhuizen GP, Wagner SJ, Kather JN. The future landscape of large language models in medicine. COMMUNICATIONS MEDICINE 2023; 3:141. [PMID: 37816837 PMCID: PMC10564921 DOI: 10.1038/s43856-023-00370-1] [Citation(s) in RCA: 247] [Impact Index Per Article: 123.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2023] [Accepted: 09/21/2023] [Indexed: 10/12/2023] Open
Abstract
Large language models (LLMs) are artificial intelligence (AI) tools specifically trained to process and generate text. LLMs attracted substantial public attention after OpenAI's ChatGPT was made publicly available in November 2022. LLMs can often answer questions, summarize, paraphrase and translate text on a level that is nearly indistinguishable from human capabilities. The possibility to actively interact with models like ChatGPT makes LLMs attractive tools in various fields, including medicine. While these models have the potential to democratize medical knowledge and facilitate access to healthcare, they could equally distribute misinformation and exacerbate scientific misconduct due to a lack of accountability and transparency. In this article, we provide a systematic and comprehensive overview of the potentials and limitations of LLMs in clinical practice, medical research and medical education.
Collapse
Affiliation(s)
- Jan Clusmann
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
- Department of Medicine III, University Hospital RWTH Aachen, Aachen, Germany
| | - Fiona R Kolbinger
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
- Department of Visceral, Thoracic and Vascular Surgery, University Hospital and Faculty of Medicine Carl Gustav Carus, TUD Dresden University of Technology, Dresden, Germany
| | - Hannah Sophie Muti
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
- Department of Visceral, Thoracic and Vascular Surgery, University Hospital and Faculty of Medicine Carl Gustav Carus, TUD Dresden University of Technology, Dresden, Germany
| | - Zunamys I Carrero
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
| | - Jan-Niklas Eckardt
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
- Department of Medicine I, University Hospital Dresden, Dresden, Germany
| | - Narmin Ghaffari Laleh
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
- Department of Medicine III, University Hospital RWTH Aachen, Aachen, Germany
| | - Chiara Maria Lavinia Löffler
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
- Department of Medicine I, University Hospital Dresden, Dresden, Germany
| | - Sophie-Caroline Schwarzkopf
- Department of Visceral, Thoracic and Vascular Surgery, University Hospital and Faculty of Medicine Carl Gustav Carus, TUD Dresden University of Technology, Dresden, Germany
| | - Michaela Unger
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
| | - Gregory P Veldhuizen
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany
| | - Sophia J Wagner
- Helmholtz Munich-German Research Center for Environment and Health, Munich, Germany
- School of Computation, Information and Technology, Technical University of Munich, Munich, Germany
| | - Jakob Nikolas Kather
- Else Kroener Fresenius Center for Digital Health, TUD Dresden University of Technology, Dresden, Germany.
- Department of Medicine III, University Hospital RWTH Aachen, Aachen, Germany.
- Department of Medicine I, University Hospital Dresden, Dresden, Germany.
- Medical Oncology, National Center for Tumor Diseases (NCT), University Hospital Heidelberg, Heidelberg, Germany.
| |
Collapse
|
139
|
Goodman RS, Patrinely JR, Stone CA, Zimmerman E, Donald RR, Chang SS, Berkowitz ST, Finn AP, Jahangir E, Scoville EA, Reese TS, Friedman DL, Bastarache JA, van der Heijden YF, Wright JJ, Ye F, Carter N, Alexander MR, Choe JH, Chastain CA, Zic JA, Horst SN, Turker I, Agarwal R, Osmundson E, Idrees K, Kiernan CM, Padmanabhan C, Bailey CE, Schlegel CE, Chambless LB, Gibson MK, Osterman TJ, Wheless LE, Johnson DB. Accuracy and Reliability of Chatbot Responses to Physician Questions. JAMA Netw Open 2023; 6:e2336483. [PMID: 37782499 PMCID: PMC10546234 DOI: 10.1001/jamanetworkopen.2023.36483] [Citation(s) in RCA: 138] [Impact Index Per Article: 69.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/16/2023] [Accepted: 08/22/2023] [Indexed: 10/03/2023] Open
Abstract
Importance Natural language processing tools, such as ChatGPT (generative pretrained transformer, hereafter referred to as chatbot), have the potential to radically enhance the accessibility of medical information for health professionals and patients. Assessing the safety and efficacy of these tools in answering physician-generated questions is critical to determining their suitability in clinical settings, facilitating complex decision-making, and optimizing health care efficiency. Objective To assess the accuracy and comprehensiveness of chatbot-generated responses to physician-developed medical queries, highlighting the reliability and limitations of artificial intelligence-generated medical information. Design, Setting, and Participants Thirty-three physicians across 17 specialties generated 284 medical questions that they subjectively classified as easy, medium, or hard with either binary (yes or no) or descriptive answers. The physicians then graded the chatbot-generated answers to these questions for accuracy (6-point Likert scale with 1 being completely incorrect and 6 being completely correct) and completeness (3-point Likert scale, with 1 being incomplete and 3 being complete plus additional context). Scores were summarized with descriptive statistics and compared using the Mann-Whitney U test or the Kruskal-Wallis test. The study (including data analysis) was conducted from January to May 2023. Main Outcomes and Measures Accuracy, completeness, and consistency over time and between 2 different versions (GPT-3.5 and GPT-4) of chatbot-generated medical responses. Results Across all questions (n = 284) generated by 33 physicians (31 faculty members and 2 recent graduates from residency or fellowship programs) across 17 specialties, the median accuracy score was 5.5 (IQR, 4.0-6.0) (between almost completely and complete correct) with a mean (SD) score of 4.8 (1.6) (between mostly and almost completely correct). The median completeness score was 3.0 (IQR, 2.0-3.0) (complete and comprehensive) with a mean (SD) score of 2.5 (0.7). For questions rated easy, medium, and hard, the median accuracy scores were 6.0 (IQR, 5.0-6.0), 5.5 (IQR, 5.0-6.0), and 5.0 (IQR, 4.0-6.0), respectively (mean [SD] scores were 5.0 [1.5], 4.7 [1.7], and 4.6 [1.6], respectively; P = .05). Accuracy scores for binary and descriptive questions were similar (median score, 6.0 [IQR, 4.0-6.0] vs 5.0 [IQR, 3.4-6.0]; mean [SD] score, 4.9 [1.6] vs 4.7 [1.6]; P = .07). Of 36 questions with scores of 1.0 to 2.0, 34 were requeried or regraded 8 to 17 days later with substantial improvement (median score 2.0 [IQR, 1.0-3.0] vs 4.0 [IQR, 2.0-5.3]; P < .01). A subset of questions, regardless of initial scores (version 3.5), were regenerated and rescored using version 4 with improvement (mean accuracy [SD] score, 5.2 [1.5] vs 5.7 [0.8]; median score, 6.0 [IQR, 5.0-6.0] for original and 6.0 [IQR, 6.0-6.0] for rescored; P = .002). Conclusions and Relevance In this cross-sectional study, chatbot generated largely accurate information to diverse medical queries as judged by academic physician specialists with improvement over time, although it had important limitations. Further research and model development are needed to correct inaccuracies and for validation.
Collapse
Affiliation(s)
| | - J. Randall Patrinely
- Department of Dermatology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Cosby A. Stone
- Department of Allergy, Pulmonology, and Critical Care, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Eli Zimmerman
- Department of Neurology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Rebecca R. Donald
- Department of Anesthesiology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Sam S. Chang
- Department of Urology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Sean T. Berkowitz
- Vanderbilt Eye Institute, Department of Ophthalmology, Vanderbilt University Medical, Nashville, Tennessee
| | - Avni P. Finn
- Vanderbilt Eye Institute, Department of Ophthalmology, Vanderbilt University Medical, Nashville, Tennessee
| | - Eiman Jahangir
- Department of Cardiovascular Medicine, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Elizabeth A. Scoville
- Department of Gastroenterology, Hepatology, and Nutrition, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Tyler S. Reese
- Department of Rheumatology and Immunology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Debra L. Friedman
- Department of Pediatric Hematology/Oncology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Julie A. Bastarache
- Department of Allergy, Pulmonology, and Critical Care, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Yuri F. van der Heijden
- Department of Infectious Disease, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Jordan J. Wright
- Department of Diabetes, Endocrinology, and Metabolism, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Fei Ye
- Department of Biostatistics, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Nicholas Carter
- Division of Trauma and Surgical Critical Care, University of Miami Miller School of Medicine, Miami, Florida
| | - Matthew R. Alexander
- Department of Cardiovascular Medicine and Clinical Pharmacology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Jennifer H. Choe
- Department of Hematology/Oncology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Cody A. Chastain
- Department of Infectious Disease, Vanderbilt University Medical Center, Nashville, Tennessee
| | - John A. Zic
- Department of Dermatology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Sara N. Horst
- Department of Gastroenterology, Hepatology, and Nutrition, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Isik Turker
- Department of Cardiology, Washington University School of Medicine in St Louis, St Louis, Missouri
| | - Rajiv Agarwal
- Department of Hematology/Oncology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Evan Osmundson
- Department of Radiation Oncology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Kamran Idrees
- Department of Surgical Oncology & Endocrine Surgery, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Colleen M. Kiernan
- Department of Surgical Oncology & Endocrine Surgery, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Chandrasekhar Padmanabhan
- Department of Surgical Oncology & Endocrine Surgery, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Christina E. Bailey
- Department of Surgical Oncology & Endocrine Surgery, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Cameron E. Schlegel
- Department of Surgical Oncology & Endocrine Surgery, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Lola B. Chambless
- Department of Neurological Surgery, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Michael K. Gibson
- Department of Hematology/Oncology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Travis J. Osterman
- Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Lee E. Wheless
- Department of Dermatology, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Douglas B. Johnson
- Department of Hematology/Oncology, Vanderbilt University Medical Center, Nashville, Tennessee
| |
Collapse
|
140
|
Benichou L. The role of using ChatGPT AI in writing medical scientific articles. JOURNAL OF STOMATOLOGY, ORAL AND MAXILLOFACIAL SURGERY 2023; 124:101456. [PMID: 36966950 DOI: 10.1016/j.jormas.2023.101456] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/15/2023] [Revised: 03/15/2023] [Accepted: 03/23/2023] [Indexed: 06/18/2023]
Abstract
The use of artificial intelligence (AI) in medical research is on the rise. This article explores the role of using ChatGPT, a language model developed by OpenAI, in writing medical scientific articles. The material and methods used included a comparative analysis of medical scientific articles produced with and without the use of ChatGPT. The results suggest that the use of ChatGPT can be a useful tool for scientists to increase the production of higher quality medical scientific articles, but it is important to note that AI cannot fully replace human authors. In conclusion, scientists should consider ChatGPT as an additional tool to produce higher quality medical scientific articles more quickly.
Collapse
Affiliation(s)
- L Benichou
- Service de chirurgie maxillo-faciale et stomatologie, Groupe Hospitalier Paris St-Joseph, 185 rue Raymond Losserand 75014 Paris, France.
| |
Collapse
|
141
|
Cai LZ, Shaheen A, Jin A, Fukui R, Yi JS, Yannuzzi N, Alabiad C. Performance of Generative Large Language Models on Ophthalmology Board-Style Questions. Am J Ophthalmol 2023; 254:141-149. [PMID: 37339728 DOI: 10.1016/j.ajo.2023.05.024] [Citation(s) in RCA: 65] [Impact Index Per Article: 32.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2023] [Revised: 05/27/2023] [Accepted: 05/28/2023] [Indexed: 06/22/2023]
Abstract
PURPOSE To investigate the ability of generative artificial intelligence models to answer ophthalmology board-style questions. DESIGN Experimental study. METHODS This study evaluated 3 large language models (LLMs) with chat interfaces, Bing Chat (Microsoft) and ChatGPT 3.5 and 4.0 (OpenAI), using 250 questions from the Basic Science and Clinical Science Self-Assessment Program. Although ChatGPT is trained on information last updated in 2021, Bing Chat incorporates a more recently indexed internet search to generate its answers. Performance was compared with human respondents. Questions were categorized by complexity and patient care phase, and instances of information fabrication or nonlogical reasoning were documented. MAIN OUTCOME MEASURES Primary outcome was response accuracy. Secondary outcomes were performance in question subcategories and hallucination frequency. RESULTS Human respondents had an average accuracy of 72.2%. ChatGPT-3.5 scored the lowest (58.8%), whereas ChatGPT-4.0 (71.6%) and Bing Chat (71.2%) performed comparably. ChatGPT-4.0 excelled in workup-type questions (odds ratio [OR], 3.89, 95% CI, 1.19-14.73, P = .03) compared with diagnostic questions, but struggled with image interpretation (OR, 0.14, 95% CI, 0.05-0.33, P < .01) when compared with single-step reasoning questions. Against single-step questions, Bing Chat also faced difficulties with image interpretation (OR, 0.18, 95% CI, 0.08-0.44, P < .01) and multi-step reasoning (OR, 0.30, 95% CI, 0.11-0.84, P = .02). ChatGPT-3.5 had the highest rate of hallucinations and nonlogical reasoning (42.4%), followed by ChatGPT-4.0 (18.0%) and Bing Chat (25.6%). CONCLUSIONS LLMs (particularly ChatGPT-4.0 and Bing Chat) can perform similarly with human respondents answering questions from the Basic Science and Clinical Science Self-Assessment Program. The frequency of hallucinations and nonlogical reasoning suggests room for improvement in the performance of conversational agents in the medical domain.
Collapse
Affiliation(s)
- Louis Z Cai
- From the Bascom Palmer Eye Institute, Miami, Florida, USA (L.Z.C., A.S., J.S.Y., N.Y., C.A.).
| | - Abdulla Shaheen
- From the Bascom Palmer Eye Institute, Miami, Florida, USA (L.Z.C., A.S., J.S.Y., N.Y., C.A.)
| | - Andrew Jin
- Yale Eye Center, New Haven, Connecticut, USA (A.J.)
| | - Riya Fukui
- Houston Rehabilitation Group, Houston, Texas, USA (R.F.)
| | - Jonathan S Yi
- From the Bascom Palmer Eye Institute, Miami, Florida, USA (L.Z.C., A.S., J.S.Y., N.Y., C.A.)
| | - Nicolas Yannuzzi
- From the Bascom Palmer Eye Institute, Miami, Florida, USA (L.Z.C., A.S., J.S.Y., N.Y., C.A.)
| | - Chrisfouad Alabiad
- From the Bascom Palmer Eye Institute, Miami, Florida, USA (L.Z.C., A.S., J.S.Y., N.Y., C.A.)
| |
Collapse
|
142
|
Krüger L, Krotsetis S, Nydahl P. [ChatGPT: curse or blessing in nursing care?]. Med Klin Intensivmed Notfmed 2023; 118:534-539. [PMID: 37401955 DOI: 10.1007/s00063-023-01038-3] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2023] [Revised: 04/28/2023] [Accepted: 06/03/2023] [Indexed: 07/05/2023]
Abstract
Artificial intelligence (AI) has been used in healthcare for some years for risk detection, diagnostics, documentation, education and training and other purposes. A new open AI application is ChatGPT, which is accessible to everyone. The application of ChatGPT as AI in education, training or studies is currently being discussed from many perspectives. It is questionable whether ChatGPT can and should also support nursing professions in health care. The aim of this review article is to show and critically discuss possible areas of application of ChatGPT in theory and practice with a focus on nursing practice, pedagogy, nursing research and nursing development.
Collapse
Affiliation(s)
- Lars Krüger
- Herz- und Diabeteszentrum NRW, Universitätsklinikum der Ruhr-Universität Bochum, Bad Oeynhausen, Deutschland
| | - Susanne Krotsetis
- Pflegeentwicklung und Pflegewissenschaft angegliedert der Pflegedirektion, des Universitätsklinikums Schleswig-Holstein, Campus Lübeck, Lübeck, Deutschland
| | - Peter Nydahl
- Pflegeforschung und -entwicklung, Pflegedirektion, Universitätsklinikum Schleswig-Holstein, Haus V40, Arnold-Heller-Str. 3, 24105, Kiel, Deutschland.
- Universitätsinstitut für Pflegewissenschaft und -praxis, Paracelsus Medizinische Privatuniversität, Salzburg, Österreich.
| |
Collapse
|
143
|
Castonguay A, Farthing P, Davies S, Vogelsang L, Kleib M, Risling T, Green N. Revolutionizing nursing education through Ai integration: A reflection on the disruptive impact of ChatGPT. NURSE EDUCATION TODAY 2023; 129:105916. [PMID: 37515957 DOI: 10.1016/j.nedt.2023.105916] [Citation(s) in RCA: 30] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/03/2023] [Revised: 06/06/2023] [Accepted: 07/17/2023] [Indexed: 07/31/2023]
Abstract
Artificial intelligence (AI) is driving global change. An AI language model like ChatGPT could revolutionize the delivery of nursing education in the future. ChatGPT is an AI-enabled text generator that has garnered significant attention due to its ability to engage in conversations and answer questions. Nurse educators play a crucial role in preparing nursing students for a technology-integrated healthcare system, and the emergence of ChatGPT presents both opportunities and challenges. While the technology has limitations and potential biases, it also has the potential to benefit students by facilitating learning, improving digital literacy, and encouraging critical thinking about AI integration in healthcare. Nurse educators can incorporate ChatGPT into their curriculum through formative or summative assessments and should prioritize faculty development to understand and use AI technologies effectively. Collaboration between educational institutions, regulatory bodies, and educators is crucial to establish provincial and national competencies and frameworks that reflect the increasing importance of AI in nursing education and practice. It is paramount that nurses and nurse educators be open to AI-enabled innovations as well as continue to critically think about their potential value to advance the profession so nurses are better prepared to lead the digital future.
Collapse
Affiliation(s)
- Alexandre Castonguay
- University of Montreal, Nursing Faculty, Marguerite-d'Youville Campus, C.P. 6128 succ. Centre-ville, Montréal, Québec H3C 3J7, Canada.
| | - Pamela Farthing
- Saskatchewan Polytechnic, Faculty, School of Nursing, SCBScN, Canada.
| | | | - Laura Vogelsang
- University of Lethbridge, Faculty of Health Sciences, Canada.
| | - Manal Kleib
- University of Alberta, Faculty of Nursing, Canada.
| | | | - Nadia Green
- University of Alberta, Faculty of Nursing, Canada.
| |
Collapse
|
144
|
Hosseini M, Resnik DB, Holmes K. The ethics of disclosing the use of artificial intelligence tools in writing scholarly manuscripts. RESEARCH ETHICS 2023; 19:449-465. [PMID: 39749232 PMCID: PMC11694804 DOI: 10.1177/17470161231180449] [Citation(s) in RCA: 29] [Impact Index Per Article: 14.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/04/2025]
Abstract
In this article, we discuss ethical issues related to using and disclosing artificial intelligence (AI) tools, such as ChatGPT and other systems based on large language models (LLMs), to write or edit scholarly manuscripts. Some journals, such as Science, have banned the use of LLMs because of the ethical problems they raise concerning responsible authorship. We argue that this is not a reasonable response to the moral conundrums created by the use of LLMs because bans are unenforceable and would encourage undisclosed use of LLMs. Furthermore, LLMs can be useful in writing, reviewing and editing text, and promote equity in science. Others have argued that LLMs should be mentioned in the acknowledgments since they do not meet all the authorship criteria. We argue that naming LLMs as authors or mentioning them in the acknowledgments are both inappropriate forms of recognition because LLMs do not have free will and therefore cannot be held morally or legally responsible for what they do. Tools in general, and software in particular, are usually cited in-text, followed by being mentioned in the references. We provide suggestions to improve APA Style for referencing ChatGPT to specifically indicate the contributor who used LLMs (because interactions are stored on personal user accounts), the used version and model (because the same version could use different language models and generate dissimilar responses, e.g., ChatGPT May 12 Version GPT3.5 or GPT4), and the time of usage (because LLMs evolve fast and generate dissimilar responses over time). We recommend that researchers who use LLMs: (1) disclose their use in the introduction or methods section to transparently describe details such as used prompts and note which parts of the text are affected, (2) use in-text citations and references (to recognize their used applications and improve findability and indexing), and (3) record and submit their relevant interactions with LLMs as supplementary material or appendices.
Collapse
Affiliation(s)
| | - David B Resnik
- National Institute of Environmental Health Sciences, USA
| | - Kristi Holmes
- Northwestern University Feinberg School of Medicine, USA
| |
Collapse
|
145
|
Diaz Milian R, Moreno Franco P, Freeman WD, Halamka JD. Revolution or Peril? The Controversial Role of Large Language Models in Medical Manuscript Writing. Mayo Clin Proc 2023; 98:1444-1448. [PMID: 37793723 DOI: 10.1016/j.mayocp.2023.07.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/23/2023] [Accepted: 07/13/2023] [Indexed: 10/06/2023]
Affiliation(s)
- Ricardo Diaz Milian
- Division of Critical Care Medicine, Mayo Clinic, Jacksonville, FL; Department of Anesthesiology, Mayo Clinic, Jacksonville, FL.
| | - Pablo Moreno Franco
- Department of Anesthesiology, Mayo Clinic, Jacksonville, FL; Department of Transplant Medicine, Mayo Clinic, Jacksonville, FL
| | - William D Freeman
- Department of Neurology and Neurosurgery, Mayo Clinic, Jacksonville, FL
| | - John D Halamka
- Department of Emergency Medicine, Mayo Clinic, Rochester, MN; Department of Internal Medicine, Mayo Clinic, Rochester, MN; Mayo Clinic Platform, Mayo Clinic, Rochester, MN
| |
Collapse
|
146
|
Ergun Y. Redefining authorship in the era of artificial intelligence: balancing ethics, transparency, and progress. ESMO Open 2023; 8:101634. [PMID: 37659291 PMCID: PMC10480051 DOI: 10.1016/j.esmoop.2023.101634] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Accepted: 08/04/2023] [Indexed: 09/04/2023] Open
Affiliation(s)
- Y Ergun
- Department of Medical Oncology, Batman World Hospital, Batman, Turkey.
| |
Collapse
|
147
|
Knox A, Bass N, Khakoo Y. You Can Run but You Can't Hide: Artificial Intelligence Is Here. Pediatr Neurol 2023; 147:163-164. [PMID: 37024352 DOI: 10.1016/j.pediatrneurol.2023.03.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 04/08/2023]
Affiliation(s)
- Andrew Knox
- Division of Pediatric Neurology, Department of Neurology, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin.
| | - Nancy Bass
- Division of Child Neurology, Children's Wisconsin, Medical College of Wisconsin, Milwaukee, Wisconsin
| | - Yasmin Khakoo
- Division of Child Neurology, Departments of Pediatrics and Neurology, Memorial Sloan Kettering Cancer Center, New York, New York
| |
Collapse
|
148
|
Grillo R, Quinta Reis BA, Melhem-Elias F. The risks and benefits of utilizing artificial intelligence in oral and maxillofacial surgery. JOURNAL OF STOMATOLOGY, ORAL AND MAXILLOFACIAL SURGERY 2023; 124:101492. [PMID: 37149261 DOI: 10.1016/j.jormas.2023.101492] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Accepted: 05/03/2023] [Indexed: 05/08/2023]
Affiliation(s)
- Ricardo Grillo
- Department of Oral and Maxillofacial Surgery, University of São Paulo School of Dentistry, São Paulo, SP, Brazil; Department of Oral and Maxillofacial Surgery, Faculdade Patos de Minas, Brasília, DF, Brazil.
| | | | - Fernando Melhem-Elias
- Department of Oral and Maxillofacial Surgery, University of São Paulo School of Dentistry, São Paulo, SP, Brazil
| |
Collapse
|
149
|
Hwang SI, Lim JS, Lee RW, Matsui Y, Iguchi T, Hiraki T, Ahn H. Is ChatGPT a "Fire of Prometheus" for Non-Native English-Speaking Researchers in Academic Writing? Korean J Radiol 2023; 24:952-959. [PMID: 37793668 PMCID: PMC10550740 DOI: 10.3348/kjr.2023.0773] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Accepted: 08/18/2023] [Indexed: 10/06/2023] Open
Abstract
Large language models (LLMs) such as ChatGPT have garnered considerable interest for their potential to aid non-native English-speaking researchers. These models can function as personal, round-the-clock English tutors, akin to how Prometheus in Greek mythology bestowed fire upon humans for their advancement. LLMs can be particularly helpful for non-native researchers in writing the Introduction and Discussion sections of manuscripts, where they often encounter challenges. However, using LLMs to generate text for research manuscripts entails concerns such as hallucination, plagiarism, and privacy issues; to mitigate these risks, authors should verify the accuracy of generated content, employ text similarity detectors, and avoid inputting sensitive information into their prompts. Consequently, it may be more prudent to utilize LLMs for editing and refining text rather than generating large portions of text. Journal policies concerning the use of LLMs vary, but transparency in disclosing artificial intelligence tool usage is emphasized. This paper aims to summarize how LLMs can lower the barrier to academic writing in English, enabling researchers to concentrate on domain-specific research, provided they are used responsibly and cautiously.
Collapse
Affiliation(s)
- Sung Il Hwang
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
- Department of Radiology, Graduate School of Medicine, Dentistry and Pharmaceutical Sciences, Okayama University, Okayama, Japan
| | - Joon Seo Lim
- Scientific Publications Team, Clinical Research Center, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Republic of Korea.
| | - Ro Woon Lee
- Department of Radiology, Inha University Hospital, Incheon, Republic of Korea
| | - Yusuke Matsui
- Department of Radiology, Faculty of Medicine, Dentistry and Pharmaceutical Sciences, Okayama University, Okayama, Japan
| | - Toshihiro Iguchi
- Department of Radiological Technology, Faculty of Health Sciences, Okayama University, Okayama, Japan
| | - Takao Hiraki
- Department of Radiology, Faculty of Medicine, Dentistry and Pharmaceutical Sciences, Okayama University, Okayama, Japan
| | - Hyungwoo Ahn
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| |
Collapse
|
150
|
Rao A, Kim J, Kamineni M, Pang M, Lie W, Dreyer KJ, Succi MD. Evaluating GPT as an Adjunct for Radiologic Decision Making: GPT-4 Versus GPT-3.5 in a Breast Imaging Pilot. J Am Coll Radiol 2023; 20:990-997. [PMID: 37356806 PMCID: PMC10733745 DOI: 10.1016/j.jacr.2023.05.003] [Citation(s) in RCA: 104] [Impact Index Per Article: 52.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Revised: 05/16/2023] [Accepted: 05/23/2023] [Indexed: 06/27/2023]
Abstract
OBJECTIVE Despite rising popularity and performance, studies evaluating the use of large language models for clinical decision support are lacking. Here, we evaluate ChatGPT (Generative Pre-trained Transformer)-3.5 and GPT-4's (OpenAI, San Francisco, California) capacity for clinical decision support in radiology via the identification of appropriate imaging services for two important clinical presentations: breast cancer screening and breast pain. METHODS We compared ChatGPT's responses to the ACR Appropriateness Criteria for breast pain and breast cancer screening. Our prompt formats included an open-ended (OE) and a select all that apply (SATA) format. Scoring criteria evaluated whether proposed imaging modalities were in accordance with ACR guidelines. Three replicate entries were conducted for each prompt, and the average of these was used to determine final scores. RESULTS Both ChatGPT-3.5 and ChatGPT-4 achieved an average OE score of 1.830 (out of 2) for breast cancer screening prompts. ChatGPT-3.5 achieved a SATA average percentage correct of 88.9%, compared with ChatGPT-4's average percentage correct of 98.4% for breast cancer screening prompts. For breast pain, ChatGPT-3.5 achieved an average OE score of 1.125 (out of 2) and a SATA average percentage correct of 58.3%, as compared with an average OE score of 1.666 (out of 2) and a SATA average percentage correct of 77.7%. DISCUSSION Our results demonstrate the eventual feasibility of using large language models like ChatGPT for radiologic decision making, with the potential to improve clinical workflow and responsible use of radiology services. More use cases and greater accuracy are necessary to evaluate and implement such tools.
Collapse
Affiliation(s)
- Arya Rao
- Harvard Medical School, Boston, Massachusetts; Medically Engineered Solutions in Healthcare, Innovation in Operations Research Center, Massachusetts General Hospital, Boston, Massachusetts
| | - John Kim
- Harvard Medical School, Boston, Massachusetts; Medically Engineered Solutions in Healthcare, Innovation in Operations Research Center, Massachusetts General Hospital, Boston, Massachusetts
| | - Meghana Kamineni
- Harvard Medical School, Boston, Massachusetts; Medically Engineered Solutions in Healthcare, Innovation in Operations Research Center, Massachusetts General Hospital, Boston, Massachusetts
| | - Michael Pang
- Harvard Medical School, Boston, Massachusetts; Medically Engineered Solutions in Healthcare, Innovation in Operations Research Center, Massachusetts General Hospital, Boston, Massachusetts
| | - Winston Lie
- Harvard Medical School, Boston, Massachusetts; Medically Engineered Solutions in Healthcare, Innovation in Operations Research Center, Massachusetts General Hospital, Boston, Massachusetts
| | - Keith J Dreyer
- Harvard Medical School, Boston, Massachusetts; Medically Engineered Solutions in Healthcare, Innovation in Operations Research Center, Massachusetts General Hospital, Boston, Massachusetts; Department of Radiology, Massachusetts General Hospital, Boston, Massachusetts; and Chief Data Science Officer and Chief Imaging Information Officer for Mass General Brigham, Boston, Massachusetts
| | - Marc D Succi
- Harvard Medical School, Boston, Massachusetts; Medically Engineered Solutions in Healthcare, Innovation in Operations Research Center and Associate Chair of Innovation & Commercialization, Mass General Brigham Enterprise Radiology; Executive Director, MESH Incubator. Massachusetts General Hospital, Boston, Massachusetts; and Department of Radiology, Massachusetts General Hospital, Boston, Massachusetts.
| |
Collapse
|