2
|
Mantena S, Johnson A, Oppezzo M, Schuetz N, Tolas A, Doijad R, Mattson CM, Lawrie A, Ramirez-Posada M, Linos E, King AC, Rodriguez F, Kim DS, Ashley EA. Fine-tuning Large Language Models in Behavioral Psychology for Scalable Physical Activity Coaching. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2025:2025.02.19.25322559. [PMID: 40034753 PMCID: PMC11875315 DOI: 10.1101/2025.02.19.25322559] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Personalized, smartphone-based coaching improves physical activity but relies on static, human-crafted messages. We introduce My Heart Counts (MHC)-Coach, a large language model fine-tuned on the Transtheoretical Model of Change. MHC-Coach generates messages tailored to an individual's psychology (their "stage of change"), providing personalized support to foster long-term physical activity behavior change. To evaluate MHC-Coach's efficacy, 632 participants compared human-expert and MHC-Coach text-based interventions encouraging physical activity. Among messages matched to an individual's stage of change, 68.0% (N=430) preferred MHC-Coach-generated messages (P < 0.001). Blinded behavioral science experts (N=2) rated MHC-Coach messages higher than human-expert messages for perceived effectiveness (4.4 vs. 2.8) and Transtheoretical Model alignment (4.1 vs. 3.5) on a 5-point Likert scale. This work demonstrates how language models can operationalize behavioral science frameworks for personalized health coaching, promoting long-term physical activity and potentially reducing cardiovascular disease risk at scale.
Collapse
Affiliation(s)
- Sriya Mantena
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Anders Johnson
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Marily Oppezzo
- Department of Epidemiology & Population Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Stanford Prevention Research Center, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Narayan Schuetz
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Department of Psychiatry and Behavioral Sciences, Stanford University, Stanford, CA, USA
- Clinical Excellence Research Center (CERC), Stanford University, Stanford, CA, USA
| | - Alexander Tolas
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Ritu Doijad
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - C. Mikael Mattson
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Allan Lawrie
- National Heart and Lung Institute, Imperial College London, London, UK
| | - Mariana Ramirez-Posada
- Department of Dermatology, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Eleni Linos
- Department of Dermatology, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Abby C. King
- Department of Epidemiology & Population Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Stanford Prevention Research Center, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Fatima Rodriguez
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Wu Tsai Human Performance Alliance, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Cardiovascular Institute, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Daniel Seung Kim
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Wu Tsai Human Performance Alliance, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Cardiovascular Institute, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Euan A. Ashley
- Division of Cardiovascular Medicine, Department of Medicine, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Wu Tsai Human Performance Alliance, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Cardiovascular Institute, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Department of Genetics, Stanford University School of Medicine, Stanford, CA, 94305, USA
- Department of Biomedical Data Science, Stanford University School of Medicine, Stanford, CA, 94305, USA
| |
Collapse
|
3
|
Ferrara E. Large Language Models for Wearable Sensor-Based Human Activity Recognition, Health Monitoring, and Behavioral Modeling: A Survey of Early Trends, Datasets, and Challenges. SENSORS (BASEL, SWITZERLAND) 2024; 24:5045. [PMID: 39124092 PMCID: PMC11314694 DOI: 10.3390/s24155045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/10/2024] [Revised: 07/29/2024] [Accepted: 07/31/2024] [Indexed: 08/12/2024]
Abstract
The proliferation of wearable technology enables the generation of vast amounts of sensor data, offering significant opportunities for advancements in health monitoring, activity recognition, and personalized medicine. However, the complexity and volume of these data present substantial challenges in data modeling and analysis, which have been addressed with approaches spanning time series modeling to deep learning techniques. The latest frontier in this domain is the adoption of large language models (LLMs), such as GPT-4 and Llama, for data analysis, modeling, understanding, and human behavior monitoring through the lens of wearable sensor data. This survey explores the current trends and challenges in applying LLMs for sensor-based human activity recognition and behavior modeling. We discuss the nature of wearable sensor data, the capabilities and limitations of LLMs in modeling them, and their integration with traditional machine learning techniques. We also identify key challenges, including data quality, computational requirements, interpretability, and privacy concerns. By examining case studies and successful applications, we highlight the potential of LLMs in enhancing the analysis and interpretation of wearable sensor data. Finally, we propose future directions for research, emphasizing the need for improved preprocessing techniques, more efficient and scalable models, and interdisciplinary collaboration. This survey aims to provide a comprehensive overview of the intersection between wearable sensor data and LLMs, offering insights into the current state and future prospects of this emerging field.
Collapse
Affiliation(s)
- Emilio Ferrara
- Thomas Lord Department of Computer Science, University of Southern California, Los Angeles, CA 90007, USA;
- Information Sciences Institute, School of Advanced Computing, University of Southern California, Los Angeles, CA 90007, USA
| |
Collapse
|
4
|
Choudhury A, Chaudhry Z. Large Language Models and User Trust: Consequence of Self-Referential Learning Loop and the Deskilling of Health Care Professionals. J Med Internet Res 2024; 26:e56764. [PMID: 38662419 PMCID: PMC11082730 DOI: 10.2196/56764] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Revised: 03/12/2024] [Accepted: 03/20/2024] [Indexed: 04/26/2024] Open
Abstract
As the health care industry increasingly embraces large language models (LLMs), understanding the consequence of this integration becomes crucial for maximizing benefits while mitigating potential pitfalls. This paper explores the evolving relationship among clinician trust in LLMs, the transition of data sources from predominantly human-generated to artificial intelligence (AI)-generated content, and the subsequent impact on the performance of LLMs and clinician competence. One of the primary concerns identified in this paper is the LLMs' self-referential learning loops, where AI-generated content feeds into the learning algorithms, threatening the diversity of the data pool, potentially entrenching biases, and reducing the efficacy of LLMs. While theoretical at this stage, this feedback loop poses a significant challenge as the integration of LLMs in health care deepens, emphasizing the need for proactive dialogue and strategic measures to ensure the safe and effective use of LLM technology. Another key takeaway from our investigation is the role of user expertise and the necessity for a discerning approach to trusting and validating LLM outputs. The paper highlights how expert users, particularly clinicians, can leverage LLMs to enhance productivity by off-loading routine tasks while maintaining a critical oversight to identify and correct potential inaccuracies in AI-generated content. This balance of trust and skepticism is vital for ensuring that LLMs augment rather than undermine the quality of patient care. We also discuss the risks associated with the deskilling of health care professionals. Frequent reliance on LLMs for critical tasks could result in a decline in health care providers' diagnostic and thinking skills, particularly affecting the training and development of future professionals. The legal and ethical considerations surrounding the deployment of LLMs in health care are also examined. We discuss the medicolegal challenges, including liability in cases of erroneous diagnoses or treatment advice generated by LLMs. The paper references recent legislative efforts, such as The Algorithmic Accountability Act of 2023, as crucial steps toward establishing a framework for the ethical and responsible use of AI-based technologies in health care. In conclusion, this paper advocates for a strategic approach to integrating LLMs into health care. By emphasizing the importance of maintaining clinician expertise, fostering critical engagement with LLM outputs, and navigating the legal and ethical landscape, we can ensure that LLMs serve as valuable tools in enhancing patient care and supporting health care professionals. This approach addresses the immediate challenges posed by integrating LLMs and sets a foundation for their maintainable and responsible use in the future.
Collapse
Affiliation(s)
- Avishek Choudhury
- Industrial and Management Systems Engineering, West Virginia University, Morgantown, WV, United States
| | - Zaira Chaudhry
- Industrial and Management Systems Engineering, West Virginia University, Morgantown, WV, United States
| |
Collapse
|