1
|
Kim N, Lee J, Park SH, On Y, Lee J, Keum M, Oh S, Song Y, Lee J, Won GH, Shin JS, Lho SK, Hwang YJ, Kim TS. GPT-4 generated psychological reports in psychodynamic perspective: a pilot study on quality, risk of hallucination and client satisfaction. Front Psychiatry 2025; 16:1473614. [PMID: 40177594 PMCID: PMC11963773 DOI: 10.3389/fpsyt.2025.1473614] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/31/2024] [Accepted: 02/25/2025] [Indexed: 04/05/2025] Open
Abstract
Background Recently, there have been active proposals on how to utilize large language models (LLMs) in the fields of psychiatry and counseling. It would be interesting to develop programs with LLMs that generate psychodynamic assessments to help individuals gain insights about themselves, and to evaluate the features of such services. However, studies on this subject are rare. This pilot study aims to evaluate quality, risk of hallucination (incorrect AI-generated information), and client satisfaction with psychodynamic psychological reports generated by GPT-4. Methods The report comprised five components: psychodynamic formulation, psychopathology, parental influence, defense mechanisms, and client strengths. Participants were recruited from individuals distressed by repetitive interpersonal issues. The study was conducted in three steps: 1) Questions provided to participants, designed to create psychodynamic formulations: 14 questions were generated by GPT for inferring psychodynamic formulations, while 6 fixed questions focused on the participants' relationship with their parents. A total of 20 questions were provided. Using participants' responses to these questions, GPT-4 generated the psychological reports. 2) Seven professors of psychiatry from different university hospitals evaluated the quality and risk of hallucinations in the psychological reports by reading the reports only, without meeting the participants. This quality assessment compared the psychological reports generated by GPT-4 with those inferred by the experts. 3) Participants evaluated their satisfaction with the psychological reports. All assessments were conducted using self-report questionnaires based on a Likert scale developed for this study. Results A total of 10 participants were recruited, and the average age was 32 years. The median response indicated that quality of all five components of the psychological report was similar to the level inferred by the experts. The risk of hallucination was assessed as ranging from unlikely to minor. According to the median response in the satisfaction evaluation, the participants agreed that the report is clearly understandable, insightful, credible, useful, satisfying, and recommendable. Conclusion This study suggests the possibility that artificial intelligence could assist users by providing psychodynamic interpretations.
Collapse
Affiliation(s)
- Namwoo Kim
- Department of Clinical Medical Sciences, Seoul National University, Seoul, Republic of Korea
| | - Jiseon Lee
- Department of Psychiatry, Uijeongbu Eulji Medical Center, Eulji University School of Medicine, Uijeongbu, Republic of Korea
| | - Sung Hyeon Park
- Department of Medical Informatics, The Catholic University of Korea College of Medicine, Seoul, Republic of Korea
| | - Yoonseo On
- College of Medicine, Ewha Women’s University, Seoul, Republic of Korea
| | - Jieun Lee
- Department of Pediatrics, Inje University College of Medicine, Ilsan Paik Hospital, Goyang, Republic of Korea
| | - Musung Keum
- Department of Neuropsychiatry, Hallym University Dongtan Sacred Heart Hospital, Gyeonggi, Republic of Korea
| | - Sanghoon Oh
- Department of Psychiatry, Uijeongbu Eulji Medical Center, Eulji University School of Medicine, Uijeongbu, Republic of Korea
| | - Yoojin Song
- Department of Psychiatry, Kangwon National University Hospital, Chuncheon, Republic of Korea
| | - Junhee Lee
- Department of Psychiatry, Seoul St. Mary’s Hospital, The Catholic University of Korea, College of Medicine, Seoul, Republic of Korea
| | - Geun Hui Won
- Department of Psychiatry, Chungnam National University Sejong Hospital, Sejong, Republic of Korea
| | - Joon Sung Shin
- Department of Neuropsychiatry, Seoul National University Hospital, Seoul, Republic of Korea
| | - Silvia Kyungjin Lho
- Department of Psychiatry, SMG-SNU Boramae Medical Center, Seoul, Republic of Korea
| | - Yoon Jung Hwang
- Department of Psychiatry, Seoul St. Mary’s Hospital, The Catholic University of Korea, College of Medicine, Seoul, Republic of Korea
| | - Tae-Suk Kim
- Department of Psychiatry, Seoul St. Mary’s Hospital, The Catholic University of Korea, College of Medicine, Seoul, Republic of Korea
| |
Collapse
|
2
|
Kolding S, Lundin RM, Hansen L, Østergaard SD. Use of generative artificial intelligence (AI) in psychiatry and mental health care: a systematic review. Acta Neuropsychiatr 2024; 37:e37. [PMID: 39523628 DOI: 10.1017/neu.2024.50] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/16/2024]
Abstract
OBJECTIVES Tools based on generative artificial intelligence (AI) such as ChatGPT have the potential to transform modern society, including the field of medicine. Due to the prominent role of language in psychiatry, e.g., for diagnostic assessment and psychotherapy, these tools may be particularly useful within this medical field. Therefore, the aim of this study was to systematically review the literature on generative AI applications in psychiatry and mental health. METHODS We conducted a systematic review following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses guidelines. The search was conducted across three databases, and the resulting articles were screened independently by two researchers. The content, themes, and findings of the articles were qualitatively assessed. RESULTS The search and screening process resulted in the inclusion of 40 studies. The median year of publication was 2023. The themes covered in the articles were mainly mental health and well-being in general - with less emphasis on specific mental disorders (substance use disorder being the most prevalent). The majority of studies were conducted as prompt experiments, with the remaining studies comprising surveys, pilot studies, and case reports. Most studies focused on models that generate language, ChatGPT in particular. CONCLUSIONS Generative AI in psychiatry and mental health is a nascent but quickly expanding field. The literature mainly focuses on applications of ChatGPT, and finds that generative AI performs well, but notes that it is limited by significant safety and ethical concerns. Future research should strive to enhance transparency of methods, use experimental designs, ensure clinical relevance, and involve users/patients in the design phase.
Collapse
Affiliation(s)
- Sara Kolding
- Department of Clinical Medicine, Aarhus University, Aarhus, Denmark
- Department of Affective Disorders, Aarhus University Hospital - Psychiatry, Aarhus, Denmark
- Center for Humanities Computing, Aarhus University, Aarhus, Denmark
| | - Robert M Lundin
- Deakin University, Institute for Mental and Physical Health and Clinical Translation (IMPACT), Geelong, VIC, Australia
- Mildura Base Public Hospital, Mental Health Services, Alcohol and Other Drugs Integrated Treatment Team, Mildura, VIC, Australia
- Barwon Health, Change to Improve Mental Health (CHIME), Mental Health Drugs and Alcohol Services, Geelong, VIC, Australia
| | - Lasse Hansen
- Department of Clinical Medicine, Aarhus University, Aarhus, Denmark
- Department of Affective Disorders, Aarhus University Hospital - Psychiatry, Aarhus, Denmark
- Center for Humanities Computing, Aarhus University, Aarhus, Denmark
| | - Søren Dinesen Østergaard
- Department of Clinical Medicine, Aarhus University, Aarhus, Denmark
- Department of Affective Disorders, Aarhus University Hospital - Psychiatry, Aarhus, Denmark
| |
Collapse
|
3
|
Demirel S, Kahraman-Gokalp E, Gündüz U. From Optimism to Concern: Unveiling Sentiments and Perceptions Surrounding ChatGPT on Twitter. INTERNATIONAL JOURNAL OF HUMAN–COMPUTER INTERACTION 2024:1-23. [DOI: 10.1080/10447318.2024.2392964] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/26/2024] [Revised: 07/10/2024] [Accepted: 08/12/2024] [Indexed: 10/28/2024]
Affiliation(s)
- Sadettin Demirel
- Department of New Media and Communication, Faculty of Communication, Uskudar University, Istanbul, Turkey
| | | | - Uğur Gündüz
- Department of Journalism, Faculty of Communication, Istanbul University, Istanbul, Turkey
| |
Collapse
|
4
|
Omar M, Soffer S, Charney AW, Landi I, Nadkarni GN, Klang E. Applications of large language models in psychiatry: a systematic review. Front Psychiatry 2024; 15:1422807. [PMID: 38979501 PMCID: PMC11228775 DOI: 10.3389/fpsyt.2024.1422807] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/24/2024] [Accepted: 06/05/2024] [Indexed: 07/10/2024] Open
Abstract
Background With their unmatched ability to interpret and engage with human language and context, large language models (LLMs) hint at the potential to bridge AI and human cognitive processes. This review explores the current application of LLMs, such as ChatGPT, in the field of psychiatry. Methods We followed PRISMA guidelines and searched through PubMed, Embase, Web of Science, and Scopus, up until March 2024. Results From 771 retrieved articles, we included 16 that directly examine LLMs' use in psychiatry. LLMs, particularly ChatGPT and GPT-4, showed diverse applications in clinical reasoning, social media, and education within psychiatry. They can assist in diagnosing mental health issues, managing depression, evaluating suicide risk, and supporting education in the field. However, our review also points out their limitations, such as difficulties with complex cases and potential underestimation of suicide risks. Conclusion Early research in psychiatry reveals LLMs' versatile applications, from diagnostic support to educational roles. Given the rapid pace of advancement, future investigations are poised to explore the extent to which these models might redefine traditional roles in mental health care.
Collapse
Affiliation(s)
- Mahmud Omar
- Faculty of Medicine, Tel-Aviv University, Tel-Aviv, Israel
| | - Shelly Soffer
- Internal Medicine B, Assuta Medical Center, Ashdod, Israel
- Ben-Gurion University of the Negev, Be'er Sheva, Israel
| | | | - Isotta Landi
- Icahn School of Medicine at Mount Sinai, New York, NY, United States
| | - Girish N Nadkarni
- Hasso Plattner Institute for Digital Health at Mount Sinai, Icahn School of Medicine at Mount Sinai, New York, NY, United States
| | - Eyal Klang
- Hasso Plattner Institute for Digital Health at Mount Sinai, Icahn School of Medicine at Mount Sinai, New York, NY, United States
| |
Collapse
|
5
|
Vahia IV. Navigating New Realities in Aging Care as Artificial Intelligence Enters Clinical Practice. Am J Geriatr Psychiatry 2024; 32:267-269. [PMID: 38218703 DOI: 10.1016/j.jagp.2024.01.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/07/2024] [Accepted: 01/08/2024] [Indexed: 01/15/2024]
Affiliation(s)
- Ipsit V Vahia
- McLean Hospital (IVV), Belmont, MA; Department of Psychiatry (IVV), Harvard Medical School, Boston, MA.
| |
Collapse
|