1
|
Chen J, Chan NY, Li CT, Chan JWY, Liu Y, Li SX, Chau SWH, Leung KS, Heng PA, Lee TMC, Li TMH, Wing YK. Multimodal digital assessment of depression with actigraphy and app in Hong Kong Chinese. Transl Psychiatry 2024; 14:150. [PMID: 38499546 PMCID: PMC10948748 DOI: 10.1038/s41398-024-02873-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Revised: 03/06/2024] [Accepted: 03/12/2024] [Indexed: 03/20/2024] Open
Abstract
There is an emerging potential for digital assessment of depression. In this study, Chinese patients with major depressive disorder (MDD) and controls underwent a week of multimodal measurement including actigraphy and app-based measures (D-MOMO) to record rest-activity, facial expression, voice, and mood states. Seven machine-learning models (Random Forest [RF], Logistic regression [LR], Support vector machine [SVM], K-Nearest Neighbors [KNN], Decision tree [DT], Naive Bayes [NB], and Artificial Neural Networks [ANN]) with leave-one-out cross-validation were applied to detect lifetime diagnosis of MDD and non-remission status. Eighty MDD subjects and 76 age- and sex-matched controls completed the actigraphy, while 61 MDD subjects and 47 controls completed the app-based assessment. MDD subjects had lower mobile time (P = 0.006), later sleep midpoint (P = 0.047) and Acrophase (P = 0.024) than controls. For app measurement, MDD subjects had more frequent brow lowering (P = 0.023), less lip corner pulling (P = 0.007), higher pause variability (P = 0.046), more frequent self-reference (P = 0.024) and negative emotion words (P = 0.002), lower articulation rate (P < 0.001) and happiness level (P < 0.001) than controls. With the fusion of all digital modalities, the predictive performance (F1-score) of ANN for a lifetime diagnosis of MDD was 0.81 and 0.70 for non-remission status when combined with the HADS-D item score, respectively. Multimodal digital measurement is a feasible diagnostic tool for depression in Chinese. A combination of multimodal measurement and machine-learning approach has enhanced the performance of digital markers in phenotyping and diagnosis of MDD.
Collapse
Affiliation(s)
- Jie Chen
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China
- Department of Psychiatry, Fujian Medical University Affiliated Fuzhou Neuropsychiatric Hospital, Fuzhou, China
| | - Ngan Yin Chan
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China
| | - Chun-Tung Li
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China
| | - Joey W Y Chan
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China
| | - Yaping Liu
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China
- Center for Sleep and Circadian Medicine, The Affiliated Brain Hospital of Guangzhou Medical University, Guangzhou, Guangdong, China
| | - Shirley Xin Li
- State Key Laboratory of Brain and Cognitive Sciences, The University of Hong Kong, Hong Kong SAR, China
- Sleep Research Clinic and Laboratory, Department of Psychology, The University of Hong Kong, Hong Kong SAR, China
| | - Steven W H Chau
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China
| | - Kwong Sak Leung
- Department of Applied Data Science, Hong Kong Shue Yan University, Hong Kong SAR, China
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Pheng-Ann Heng
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Tatia M C Lee
- State Key Laboratory of Brain and Cognitive Sciences, The University of Hong Kong, Hong Kong SAR, China
- Department of Psychology, The University of Hong Kong, Hong Kong SAR, China
| | - Tim M H Li
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China.
| | - Yun-Kwok Wing
- Li Chiu Kong Family Sleep Assessment Unit, Department of Psychiatry, Faculty of Medicine, The Chinese University of Hong Kong, Shatin, Hong Kong SAR, China.
| |
Collapse
|
2
|
Khoo LS, Lim MK, Chong CY, McNaney R. Machine Learning for Multimodal Mental Health Detection: A Systematic Review of Passive Sensing Approaches. SENSORS (BASEL, SWITZERLAND) 2024; 24:348. [PMID: 38257440 PMCID: PMC10820860 DOI: 10.3390/s24020348] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Revised: 12/14/2023] [Accepted: 12/18/2023] [Indexed: 01/24/2024]
Abstract
As mental health (MH) disorders become increasingly prevalent, their multifaceted symptoms and comorbidities with other conditions introduce complexity to diagnosis, posing a risk of underdiagnosis. While machine learning (ML) has been explored to mitigate these challenges, we hypothesized that multiple data modalities support more comprehensive detection and that non-intrusive collection approaches better capture natural behaviors. To understand the current trends, we systematically reviewed 184 studies to assess feature extraction, feature fusion, and ML methodologies applied to detect MH disorders from passively sensed multimodal data, including audio and video recordings, social media, smartphones, and wearable devices. Our findings revealed varying correlations of modality-specific features in individualized contexts, potentially influenced by demographics and personalities. We also observed the growing adoption of neural network architectures for model-level fusion and as ML algorithms, which have demonstrated promising efficacy in handling high-dimensional features while modeling within and cross-modality relationships. This work provides future researchers with a clear taxonomy of methodological approaches to multimodal detection of MH disorders to inspire future methodological advancements. The comprehensive analysis also guides and supports future researchers in making informed decisions to select an optimal data source that aligns with specific use cases based on the MH disorder of interest.
Collapse
Affiliation(s)
- Lin Sze Khoo
- Department of Human-Centered Computing, Faculty of Information Technology, Monash University, Clayton, VIC 3800, Australia;
| | - Mei Kuan Lim
- School of Information Technology, Monash University Malaysia, Subang Jaya 46150, Malaysia; (M.K.L.); (C.Y.C.)
| | - Chun Yong Chong
- School of Information Technology, Monash University Malaysia, Subang Jaya 46150, Malaysia; (M.K.L.); (C.Y.C.)
| | - Roisin McNaney
- Department of Human-Centered Computing, Faculty of Information Technology, Monash University, Clayton, VIC 3800, Australia;
| |
Collapse
|
3
|
Sharma D, Singh J, Shah B, Ali F, AlZubi AA, AlZubi MA. Public mental health through social media in the post COVID-19 era. Front Public Health 2023; 11:1323922. [PMID: 38146469 PMCID: PMC10749364 DOI: 10.3389/fpubh.2023.1323922] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Accepted: 11/22/2023] [Indexed: 12/27/2023] Open
Abstract
Social media is a powerful communication tool and a reflection of our digital environment. Social media acted as an augmenter and influencer during and after COVID-19. Many of the people sharing social media posts were not actually aware of their mental health status. This situation warrants to automate the detection of mental disorders. This paper presents a methodology for the detection of mental disorders using micro facial expressions. Micro-expressions are momentary, involuntary facial expressions that can be indicative of deeper feelings and mental states. Nevertheless, manually detecting and interpreting micro-expressions can be rather challenging. A deep learning HybridMicroNet model, based on convolution neural networks, is proposed for emotion recognition from micro-expressions. Further, a case study for the detection of mental health has been undertaken. The findings demonstrated that the proposed model achieved a high accuracy when attempting to diagnose mental health disorders based on micro-expressions. The attained accuracy on the CASME dataset was 99.08%, whereas the accuracy that was achieved on SAMM dataset was 97.62%. Based on these findings, deep learning may prove to be an effective method for diagnosing mental health conditions by analyzing micro-expressions.
Collapse
Affiliation(s)
- Deepika Sharma
- Chitkara University Institute of Engineering and Technology, Chitkara University, Punjab, India
| | - Jaiteg Singh
- Chitkara University Institute of Engineering and Technology, Chitkara University, Punjab, India
| | - Babar Shah
- College of Technological Innovation, Zayed University, Dubai, United Arab Emirates
| | - Farman Ali
- Department of Computer Science and Engineering, School of Convergence, College of Computing and Informatics, Sungkyunkwan University, Seoul, Republic of Korea
| | - Ahmad Ali AlZubi
- Department of Computer Science, Community College, King Saud University, Riyadh, Saudi Arabia
| | - Mallak Ahmad AlZubi
- Faculty of Medicine, Jordan University of Science and Technology, Irbid, Jordan
| |
Collapse
|
4
|
Chen T, Hong R, Guo Y, Hao S, Hu B. MS²-GNN: Exploring GNN-Based Multimodal Fusion Network for Depression Detection. IEEE TRANSACTIONS ON CYBERNETICS 2023; 53:7749-7759. [PMID: 36194716 DOI: 10.1109/tcyb.2022.3197127] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Major depressive disorder (MDD) is one of the most common and severe mental illnesses, posing a huge burden on society and families. Recently, some multimodal methods have been proposed to learn a multimodal embedding for MDD detection and achieved promising performance. However, these methods ignore the heterogeneity/homogeneity among various modalities. Besides, earlier attempts ignore interclass separability and intraclass compactness. Inspired by the above observations, we propose a graph neural network (GNN)-based multimodal fusion strategy named modal-shared modal-specific GNN, which investigates the heterogeneity/homogeneity among various psychophysiological modalities as well as explores the potential relationship between subjects. Specifically, we develop a modal-shared and modal-specific GNN architecture to extract the inter/intramodal characteristics. Furthermore, a reconstruction network is employed to ensure fidelity within the individual modality. Moreover, we impose an attention mechanism on various embeddings to obtain a multimodal compact representation for the subsequent MDD detection task. We conduct extensive experiments on two public depression datasets and the favorable results demonstrate the effectiveness of the proposed algorithm.
Collapse
|
5
|
Hu B, Tao Y, Yang M. Detecting depression based on facial cues elicited by emotional stimuli in video. Comput Biol Med 2023; 165:107457. [PMID: 37708718 DOI: 10.1016/j.compbiomed.2023.107457] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Revised: 08/11/2023] [Accepted: 09/04/2023] [Indexed: 09/16/2023]
Abstract
Recently, depression research has received considerable attention and there is an urgent need for objective and validated methods to detect depression. Depression detection based on facial expressions may be a promising adjunct to depression detection due to its non-contact nature. Stimulated facial expressions may contain more information that is useful in detecting depression than natural facial expressions. To explore facial cues in healthy controls and depressed patients in response to different emotional stimuli, facial expressions of 62 subjects were collected while watching video stimuli, and a local face reorganization method for depression detection is proposed. The method extracts the local phase pattern features, facial action unit (AU) features and head motion features of a local face reconstructed according to facial proportions, and then fed into the classifier for classification. The classification accuracy was 76.25%, with a recall of 80.44% and a specificity of 83.21%. The results demonstrated that the negative video stimuli in the single-attribute stimulus analysis were more effective in eliciting changes in facial expressions in both healthy controls and depressed patients. Fusion of facial features under both neutral and negative stimuli was found to be useful in discriminating between healthy controls and depressed individuals. The Pearson correlation coefficient (PCC) showed that changes in the emotional stimulus paradigm were more strongly correlated with changes in subjects' facial AU when exposed to negative stimuli compared to stimuli of other attributes. These results demonstrate the feasibility of our proposed method and provide a framework for future work in assisting diagnosis.
Collapse
Affiliation(s)
- Bin Hu
- Gansu Provincial Key Laboratory of Wearable Computin, Lanzhou University, Lanzhou, 730000, Gansu, China.
| | - Yongfeng Tao
- Gansu Provincial Key Laboratory of Wearable Computin, Lanzhou University, Lanzhou, 730000, Gansu, China.
| | - Minqiang Yang
- Gansu Provincial Key Laboratory of Wearable Computin, Lanzhou University, Lanzhou, 730000, Gansu, China.
| |
Collapse
|
6
|
Xiao J, Provenza NR, Asfouri J, Myers J, Mathura RK, Metzger B, Adkinson JA, Allawala AB, Pirtle V, Oswalt D, Shofty B, Robinson ME, Mathew SJ, Goodman WK, Pouratian N, Schrater PR, Patel AB, Tolias AS, Bijanki KR, Pitkow X, Sheth SA. Decoding Depression Severity From Intracranial Neural Activity. Biol Psychiatry 2023; 94:445-453. [PMID: 36736418 PMCID: PMC10394110 DOI: 10.1016/j.biopsych.2023.01.020] [Citation(s) in RCA: 11] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/14/2022] [Revised: 01/09/2023] [Accepted: 01/25/2023] [Indexed: 02/05/2023]
Abstract
BACKGROUND Disorders of mood and cognition are prevalent, disabling, and notoriously difficult to treat. Fueling this challenge in treatment is a significant gap in our understanding of their neurophysiological basis. METHODS We recorded high-density neural activity from intracranial electrodes implanted in depression-relevant prefrontal cortical regions in 3 human subjects with severe depression. Neural recordings were labeled with depression severity scores across a wide dynamic range using an adaptive assessment that allowed sampling with a temporal frequency greater than that possible with typical rating scales. We modeled these data using regularized regression techniques with region selection to decode depression severity from the prefrontal recordings. RESULTS Across prefrontal regions, we found that reduced depression severity is associated with decreased low-frequency neural activity and increased high-frequency activity. When constraining our model to decode using a single region, spectral changes in the anterior cingulate cortex best predicted depression severity in all 3 subjects. Relaxing this constraint revealed unique, individual-specific sets of spatiospectral features predictive of symptom severity, reflecting the heterogeneous nature of depression. CONCLUSIONS The ability to decode depression severity from neural activity increases our fundamental understanding of how depression manifests in the human brain and provides a target neural signature for personalized neuromodulation therapies.
Collapse
Affiliation(s)
- Jiayang Xiao
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas; Department of Neuroscience, Baylor College of Medicine, Houston, Texas
| | - Nicole R Provenza
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Joseph Asfouri
- Department of Electrical and Computer Engineering, Rice University, Houston, Texas
| | - John Myers
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Raissa K Mathura
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Brian Metzger
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Joshua A Adkinson
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | | | - Victoria Pirtle
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Denise Oswalt
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Ben Shofty
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Meghan E Robinson
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Sanjay J Mathew
- Department of Psychiatry, Baylor College of Medicine, Houston, Texas
| | - Wayne K Goodman
- Department of Psychiatry, Baylor College of Medicine, Houston, Texas
| | - Nader Pouratian
- Department of Neurological Surgery, UT Southwestern Medical Center, Dallas, Texas
| | - Paul R Schrater
- Department of Computer Science and Engineering, University of Minnesota, Minneapolis, Minnesota; Department of Psychology, University of Minnesota, Minneapolis, Minnesota
| | - Ankit B Patel
- Department of Neuroscience, Baylor College of Medicine, Houston, Texas; Department of Electrical and Computer Engineering, Rice University, Houston, Texas; Center for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, Texas
| | - Andreas S Tolias
- Department of Neuroscience, Baylor College of Medicine, Houston, Texas; Department of Electrical and Computer Engineering, Rice University, Houston, Texas; Center for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, Texas
| | - Kelly R Bijanki
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas
| | - Xaq Pitkow
- Department of Neuroscience, Baylor College of Medicine, Houston, Texas; Department of Electrical and Computer Engineering, Rice University, Houston, Texas; Center for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, Texas
| | - Sameer A Sheth
- Department of Neurosurgery, Baylor College of Medicine, Houston, Texas.
| |
Collapse
|
7
|
Maji B, Roy AK, Nasreen S, Guha R, Routray A, Majumdar D. A Novel Technique for Detecting Depressive Disorder: A Speech Database-Based Approach. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2023; 2023:1-4. [PMID: 38082943 DOI: 10.1109/embc40787.2023.10341118] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/18/2023]
Abstract
Depression is the second most diagnosed disease in the world and is predicted to be the highest by the year 2030. Depressive disorder impacts both on mentally and physically, thus diagnosing this disorder in early stage is essential. Automatic Depression Detection (ADD) system via speech can greatly facilitate early-stage depression diagnosis. Development of such systems demands a standard balanced database. In this work, we present a novel labeled audio distress interview database. To our knowledge, this is the first depression database in Bengali language that contains audio responses from depressed and non-depressed subjects. Alongside this, we present a set of hand-crafted acoustic features that effectively detect depression mood using speech signals. Finally, we justify the quality of our developed database and the efficacy of the feature set in predicting depression using a baseline machine learning (ML) model. We believe that the annotated database will be a valuable resource for use by treating clinicians.Clinical Relevance-This research reports a new speech database in Bengali language for depression detection. This database can be used in healthcare by developing an automatic prediction model for depression detection.
Collapse
|
8
|
Ettore E, Müller P, Hinze J, Benoit M, Giordana B, Postin D, Lecomte A, Lindsay H, Robert P, König A. Digital Phenotyping for Differential Diagnosis of Major Depressive Episode: Narrative Review. JMIR Ment Health 2023; 10:e37225. [PMID: 36689265 PMCID: PMC9903183 DOI: 10.2196/37225] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/11/2022] [Revised: 09/02/2022] [Accepted: 09/30/2022] [Indexed: 01/25/2023] Open
Abstract
BACKGROUND Major depressive episode (MDE) is a common clinical syndrome. It can be found in different pathologies such as major depressive disorder (MDD), bipolar disorder (BD), posttraumatic stress disorder (PTSD), or even occur in the context of psychological trauma. However, only 1 syndrome is described in international classifications (Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition [DSM-5]/International Classification of Diseases 11th Revision [ICD-11]), which do not take into account the underlying pathology at the origin of the MDE. Clinical interviews are currently the best source of information to obtain the etiological diagnosis of MDE. Nevertheless, it does not allow an early diagnosis and there are no objective measures of extracted clinical information. To remedy this, the use of digital tools and their correlation with clinical symptomatology could be useful. OBJECTIVE We aimed to review the current application of digital tools for MDE diagnosis while highlighting shortcomings for further research. In addition, our work was focused on digital devices easy to use during clinical interview and mental health issues where depression is common. METHODS We conducted a narrative review of the use of digital tools during clinical interviews for MDE by searching papers published in PubMed/MEDLINE, Web of Science, and Google Scholar databases since February 2010. The search was conducted from June to September 2021. Potentially relevant papers were then compared against a checklist for relevance and reviewed independently for inclusion, with focus on 4 allocated topics of (1) automated voice analysis, behavior analysis by (2) video and physiological measures, (3) heart rate variability (HRV), and (4) electrodermal activity (EDA). For this purpose, we were interested in 4 frequently found clinical conditions in which MDE can occur: (1) MDD, (2) BD, (3) PTSD, and (4) psychological trauma. RESULTS A total of 74 relevant papers on the subject were qualitatively analyzed and the information was synthesized. Thus, a digital phenotype of MDE seems to emerge consisting of modifications in speech features (namely, temporal, prosodic, spectral, source, and formants) and in speech content, modifications in nonverbal behavior (head, hand, body and eyes movement, facial expressivity, and gaze), and a decrease in physiological measurements (HRV and EDA). We not only found similarities but also differences when MDE occurs in MDD, BD, PTSD, or psychological trauma. However, comparative studies were rare in BD or PTSD conditions, which does not allow us to identify clear and distinct digital phenotypes. CONCLUSIONS Our search identified markers from several modalities that hold promise for helping with a more objective diagnosis of MDE. To validate their potential, further longitudinal and prospective studies are needed.
Collapse
Affiliation(s)
- Eric Ettore
- Department of Psychiatry and Memory Clinic, University Hospital of Nice, Nice, France
| | - Philipp Müller
- Research Department Cognitive Assistants, Deutsches Forschungszentrum für Künstliche Intelligenz GmbH, Saarbrücken, Germany
| | - Jonas Hinze
- Department of Psychiatry and Psychotherapy, Saarland University Medical Center, Hombourg, Germany
| | - Michel Benoit
- Department of Psychiatry, Hopital Pasteur, University Hospital of Nice, Nice, France
| | - Bruno Giordana
- Department of Psychiatry, Hopital Pasteur, University Hospital of Nice, Nice, France
| | - Danilo Postin
- Department of Psychiatry, School of Medicine and Health Sciences, Carl von Ossietzky University of Oldenburg, Bad Zwischenahn, Germany
| | - Amandine Lecomte
- Research Department Sémagramme Team, Institut national de recherche en informatique et en automatique, Nancy, France
| | - Hali Lindsay
- Research Department Cognitive Assistants, Deutsches Forschungszentrum für Künstliche Intelligenz GmbH, Saarbrücken, Germany
| | - Philippe Robert
- Research Department, Cognition-Behaviour-Technology Lab, University Côte d'Azur, Nice, France
| | - Alexandra König
- Research Department Stars Team, Institut national de recherche en informatique et en automatique, Sophia Antipolis - Valbonne, France
| |
Collapse
|
9
|
Li N, Feng L, Hu J, Jiang L, Wang J, Han J, Gan L, He Z, Wang G. Using deeply time-series semantics to assess depressive symptoms based on clinical interview speech. Front Psychiatry 2023; 14:1104190. [PMID: 36865077 PMCID: PMC9971220 DOI: 10.3389/fpsyt.2023.1104190] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/21/2022] [Accepted: 01/31/2023] [Indexed: 02/18/2023] Open
Abstract
INTRODUCTION Depression is an affective disorder that contributes to a significant global burden of disease. Measurement-Based Care (MBC) is advocated during the full course management, with symptom assessment being an important component. Rating scales are widely used as convenient and powerful assessment tool, but they are influenced by the subjectivity and consistency of the raters. The assessment of depressive symptoms is usually conducted with a clear purpose and restricted content, such as clinical interviews based on the Hamilton Depression Rating Scale (HAMD), so that the results are easy to obtain and quantify. Artificial Intelligence (AI) techniques are used due to their objective, stable and consistent performance, and are suitable for assessing depressive symptoms. Therefore, this study applied Deep Learning (DL)-based Natural Language Processing (NLP) techniques to assess depressive symptoms during clinical interviews; thus, we proposed an algorithm model, explored the feasibility of the techniques, and evaluated their performance. METHODS The study included 329 patients with Major Depressive Episode. Clinical interviews based on the HAMD-17 were conducted by trained psychiatrists, whose speech was simultaneously recorded. A total of 387 audio recordings were included in the final analysis. A deeply time-series semantics model for the assessment of depressive symptoms based on multi-granularity and multi-task joint training (MGMT) is proposed. RESULTS The performance of MGMT is acceptable for assessing depressive symptoms with an F1 score (a metric of model performance, the harmonic mean of precision and recall) of 0.719 in classifying the four-level severity of depression and an F1 score of 0.890 in identifying the presence of depressive symptoms. DISSCUSSION This study demonstrates the feasibility of the DL and the NLP techniques applied to the clinical interview and the assessment of depressive symptoms. However, there are limitations to this study, including the lack of adequate samples, and the fact that using speech content alone to assess depressive symptoms loses the information gained through observation. A multi-dimensional model combing semantics with speech voice, facial expression, and other valuable information, as well as taking into account personalized information, is a possible direction in the future.
Collapse
Affiliation(s)
- Nanxi Li
- Beijing Key Laboratory of Mental Disorders, National Clinical Research Center for Mental Disorders and National Center for Mental Disorders, Beijing Anding Hospital, Capital Medical University, Beijing, China
| | - Lei Feng
- Beijing Key Laboratory of Mental Disorders, National Clinical Research Center for Mental Disorders and National Center for Mental Disorders, Beijing Anding Hospital, Capital Medical University, Beijing, China.,Advanced Innovation Center for Human Brain Protection, Capital Medical University, Beijing, China
| | - Jiaxue Hu
- Anhui iFLYTEK Health Co., Ltd., Hefei, China
| | - Lei Jiang
- Anhui iFLYTEK Health Co., Ltd., Hefei, China
| | - Jing Wang
- Anhui iFLYTEK Health Co., Ltd., Hefei, China
| | - Jiali Han
- Beijing Key Laboratory of Mental Disorders, National Clinical Research Center for Mental Disorders and National Center for Mental Disorders, Beijing Anding Hospital, Capital Medical University, Beijing, China
| | - Lu Gan
- Anhui iFLYTEK Health Co., Ltd., Hefei, China
| | - Zhiyang He
- Anhui iFLYTEK Health Co., Ltd., Hefei, China
| | - Gang Wang
- Beijing Key Laboratory of Mental Disorders, National Clinical Research Center for Mental Disorders and National Center for Mental Disorders, Beijing Anding Hospital, Capital Medical University, Beijing, China.,Advanced Innovation Center for Human Brain Protection, Capital Medical University, Beijing, China
| |
Collapse
|
10
|
Developing a Multimodal Monitoring System for Geriatric Depression: A Feasibility Study. COMPUTERS, INFORMATICS, NURSING : CIN 2023; 41:46-56. [PMID: 36634234 DOI: 10.1097/cin.0000000000000925] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
Abstract
The Internet of Medical Things is promising for monitoring depression symptoms. Therefore, it is necessary to develop multimodal monitoring systems tailored for elderly individuals with high feasibility and usability for further research and practice. This study comprised two phases: (1) methodological development of the system; and (2) system validation to evaluate its feasibility. We developed a system that includes a smartphone for facial and verbal expressions, a smartwatch for activity and heart rate monitoring, and an ecological momentary assessment application. A sample of 21 older Koreans aged 65 years and more was recruited from a community center. The 4-week data were collected for each participant (n = 19) using self-report questionnaires, wearable devices, and interviews and were analyzed using mixed methods. The depressive group (n = 6) indicated lower user acceptance relative to the nondepressive group (n = 13). Both groups experienced positive emotions, had regular life patterns, increased their self-interest, and stated that a system could disturb their daily activities. However, they were interested in learning new technologies and actively monitored their mental health status. Our multimodal monitoring system shows potential as a feasible and useful measure for acquiring mental health information about geriatric depression.
Collapse
|
11
|
Liu D, Liu B, Lin T, Liu G, Yang G, Qi D, Qiu Y, Lu Y, Yuan Q, Shuai SC, Li X, Liu O, Tang X, Shuai J, Cao Y, Lin H. Measuring depression severity based on facial expression and body movement using deep convolutional neural network. Front Psychiatry 2022; 13:1017064. [PMID: 36620657 PMCID: PMC9810804 DOI: 10.3389/fpsyt.2022.1017064] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/22/2022] [Accepted: 12/02/2022] [Indexed: 12/24/2022] Open
Abstract
Introduction Real-time evaluations of the severity of depressive symptoms are of great significance for the diagnosis and treatment of patients with major depressive disorder (MDD). In clinical practice, the evaluation approaches are mainly based on psychological scales and doctor-patient interviews, which are time-consuming and labor-intensive. Also, the accuracy of results mainly depends on the subjective judgment of the clinician. With the development of artificial intelligence (AI) technology, more and more machine learning methods are used to diagnose depression by appearance characteristics. Most of the previous research focused on the study of single-modal data; however, in recent years, many studies have shown that multi-modal data has better prediction performance than single-modal data. This study aimed to develop a measurement of depression severity from expression and action features and to assess its validity among the patients with MDD. Methods We proposed a multi-modal deep convolutional neural network (CNN) to evaluate the severity of depressive symptoms in real-time, which was based on the detection of patients' facial expression and body movement from videos captured by ordinary cameras. We established behavioral depression degree (BDD) metrics, which combines expression entropy and action entropy to measure the depression severity of MDD patients. Results We found that the information extracted from different modes, when integrated in appropriate proportions, can significantly improve the accuracy of the evaluation, which has not been reported in previous studies. This method presented an over 74% Pearson similarity between BDD and self-rating depression scale (SDS), self-rating anxiety scale (SAS), and Hamilton depression scale (HAMD). In addition, we tracked and evaluated the changes of BDD in patients at different stages of a course of treatment and the results obtained were in agreement with the evaluation from the scales. Discussion The BDD can effectively measure the current state of patients' depression and its changing trend according to the patient's expression and action features. Our model may provide an automatic auxiliary tool for the diagnosis and treatment of MDD.
Collapse
Affiliation(s)
- Dongdong Liu
- Department of Physics, Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen, China
| | - Bowen Liu
- Department of Psychiatry, National Clinical Research Center for Mental Disorders, The Second Xiangya Hospital of Central South University, Changsha, China
- Department of Psychiatry, Baoan Mental Health Center, Shenzhen Baoan Center for Chronic Disease Control, Shenzhen, China
| | - Tao Lin
- Oujiang Laboratory (Zhejiang Lab for Regenerative Medicine, Vision and Brain Health), Wenzhou Key Laboratory of Biophysics, Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou, Zhejiang, China
| | - Guangya Liu
- Integrated Chinese and Western Therapy of Depression Ward, Hunan Brain Hospital, Changsha, China
| | - Guoyu Yang
- Department of Physics, Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen, China
| | - Dezhen Qi
- Department of Physics, Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen, China
| | - Ye Qiu
- Department of Physics, Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen, China
- Oujiang Laboratory (Zhejiang Lab for Regenerative Medicine, Vision and Brain Health), Wenzhou Key Laboratory of Biophysics, Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou, Zhejiang, China
| | - Yuer Lu
- Department of Physics, Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen, China
- Oujiang Laboratory (Zhejiang Lab for Regenerative Medicine, Vision and Brain Health), Wenzhou Key Laboratory of Biophysics, Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou, Zhejiang, China
| | - Qinmei Yuan
- Department of Psychiatry, National Clinical Research Center for Mental Disorders, The Second Xiangya Hospital of Central South University, Changsha, China
| | - Stella C. Shuai
- Department of Biological Sciences, Northwestern University, Evanston, IL, United States
| | - Xiang Li
- Department of Physics, Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen, China
| | - Ou Liu
- Oujiang Laboratory (Zhejiang Lab for Regenerative Medicine, Vision and Brain Health), Wenzhou Key Laboratory of Biophysics, Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou, Zhejiang, China
| | - Xiangdong Tang
- Sleep Medicine Center, Mental Health Center, Department of Respiratory and Critical Care Medicine, State Key Laboratory of Biotherapy, West China Hospital, Sichuan University, Chengdu, China
| | - Jianwei Shuai
- Department of Physics, Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen, China
- Oujiang Laboratory (Zhejiang Lab for Regenerative Medicine, Vision and Brain Health), Wenzhou Key Laboratory of Biophysics, Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou, Zhejiang, China
- State Key Laboratory of Cellular Stress Biology, Innovation Center for Cell Signaling Network, National Institute for Data Science in Health and Medicine, Xiamen University, Xiamen, China
| | - Yuping Cao
- Department of Psychiatry, National Clinical Research Center for Mental Disorders, The Second Xiangya Hospital of Central South University, Changsha, China
| | - Hai Lin
- Oujiang Laboratory (Zhejiang Lab for Regenerative Medicine, Vision and Brain Health), Wenzhou Key Laboratory of Biophysics, Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou, Zhejiang, China
| |
Collapse
|
12
|
Rodriguez SN, Gullapalli AR, Maurer JM, Tirrell PS, Egala U, Anderson NE, Harenski CL, Kiehl KA. Quantitative Head Dynamics Associated with Interpersonal (Grandiose-Manipulative) Psychopathic Traits in Incarcerated Youth. JOURNAL OF PSYCHOPATHOLOGY AND BEHAVIORAL ASSESSMENT 2022; 44:1054-1063. [PMID: 37008299 PMCID: PMC10065468 DOI: 10.1007/s10862-022-09988-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/11/2022] [Indexed: 10/16/2022]
Abstract
Clinicians have long noted that individuals with elevated psychopathic traits can be characterized by unique interpersonal styles, including prolonged eye contact, invasion of interpersonal space, and frequent use of hand gestures. Such forms of nonverbal communication can be measured via hand, body, and head position and dynamics. Previous studies have developed an automated algorithm designed to capture head position and dynamics from digital recordings of clinical interviews in a sample of incarcerated adult men. We observed that higher psychopathy scores were associated with stationary head dwell time. Here, we applied a similar automated algorithm to assess head position and dynamics on videotaped clinical interviews assessing psychopathic traits from n = 242 youth housed at a maximum-security juvenile correctional facility. We observed that higher psychopathy scores (assessed via the Hare Psychopathy Checklist: Youth Version [PCL:YV]) were associated with unique patterns of head dynamics. Specifically, PCL:YV Total, Factor 1 (measuring grandiose-manipulative and callous-unemotional traits), and Facet 1 (measuring grandiose-manipulative traits) scores were associated with a higher proportion of time spent in a head dynamics pattern consisting of moderate movement away from the average head position. This study lays the groundwork for future investigations to apply quantitative methods to better understand patterns of nonverbal communication styles in clinical populations characterized by severe antisocial behavior.
Collapse
Affiliation(s)
- Samantha N. Rodriguez
- University of New Mexico, Department of Psychology, Albuquerque, NM, USA
- The Mind Research Network, Albuquerque, NM, USA
| | | | | | - Palmer S. Tirrell
- University of New Mexico, Department of Psychology, Albuquerque, NM, USA
| | - Ugesh Egala
- University of New Mexico, Department of Psychology, Albuquerque, NM, USA
| | | | | | - Kent A. Kiehl
- University of New Mexico, Department of Psychology, Albuquerque, NM, USA
- The Mind Research Network, Albuquerque, NM, USA
| |
Collapse
|
13
|
Chen ZS, Kulkarni P(P, Galatzer-Levy IR, Bigio B, Nasca C, Zhang Y. Modern views of machine learning for precision psychiatry. PATTERNS (NEW YORK, N.Y.) 2022; 3:100602. [PMID: 36419447 PMCID: PMC9676543 DOI: 10.1016/j.patter.2022.100602] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
In light of the National Institute of Mental Health (NIMH)'s Research Domain Criteria (RDoC), the advent of functional neuroimaging, novel technologies and methods provide new opportunities to develop precise and personalized prognosis and diagnosis of mental disorders. Machine learning (ML) and artificial intelligence (AI) technologies are playing an increasingly critical role in the new era of precision psychiatry. Combining ML/AI with neuromodulation technologies can potentially provide explainable solutions in clinical practice and effective therapeutic treatment. Advanced wearable and mobile technologies also call for the new role of ML/AI for digital phenotyping in mobile mental health. In this review, we provide a comprehensive review of ML methodologies and applications by combining neuroimaging, neuromodulation, and advanced mobile technologies in psychiatry practice. We further review the role of ML in molecular phenotyping and cross-species biomarker identification in precision psychiatry. We also discuss explainable AI (XAI) and neuromodulation in a closed human-in-the-loop manner and highlight the ML potential in multi-media information extraction and multi-modal data fusion. Finally, we discuss conceptual and practical challenges in precision psychiatry and highlight ML opportunities in future research.
Collapse
Affiliation(s)
- Zhe Sage Chen
- Department of Psychiatry, New York University Grossman School of Medicine, New York, NY 10016, USA
- Department of Neuroscience and Physiology, New York University Grossman School of Medicine, New York, NY 10016, USA
- The Neuroscience Institute, New York University Grossman School of Medicine, New York, NY 10016, USA
- Department of Biomedical Engineering, New York University Tandon School of Engineering, Brooklyn, NY 11201, USA
| | | | - Isaac R. Galatzer-Levy
- Department of Psychiatry, New York University Grossman School of Medicine, New York, NY 10016, USA
- Meta Reality Lab, New York, NY, USA
| | - Benedetta Bigio
- Department of Psychiatry, New York University Grossman School of Medicine, New York, NY 10016, USA
| | - Carla Nasca
- Department of Psychiatry, New York University Grossman School of Medicine, New York, NY 10016, USA
- The Neuroscience Institute, New York University Grossman School of Medicine, New York, NY 10016, USA
| | - Yu Zhang
- Department of Bioengineering, Lehigh University, Bethlehem, PA 18015, USA
- Department of Electrical and Computer Engineering, Lehigh University, Bethlehem, PA 18015, USA
| |
Collapse
|
14
|
Kaur B, Rathi S, Agrawal RK. Enhanced depression detection from speech using Quantum Whale Optimization Algorithm for feature selection. Comput Biol Med 2022; 150:106122. [PMID: 36182759 DOI: 10.1016/j.compbiomed.2022.106122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2022] [Revised: 08/27/2022] [Accepted: 09/17/2022] [Indexed: 11/03/2022]
Abstract
There is an urgent need to detect depression using a non-intrusive approach that is reliable and accurate. In this paper, a simple and efficient unimodal depression detection approach based on speech is proposed, which is non-invasive, cost-effective and computationally inexpensive. A set of spectral, temporal and spectro-temporal features is derived from the speech signal of healthy and depressed subjects. To select a minimal subset of the relevant and non-redundant speech features to detect depression, a two-phase approach based on the nature-inspired wrapper-based feature selection Quantum-based Whale Optimization Algorithm (QWOA) is proposed. Experiments are performed on the publicly available Distress Analysis Interview Corpus Wizard-of-Oz (DAICWOZ) dataset and compared with three established univariate filtering techniques for feature selection and four well-known evolutionary algorithms. The proposed model outperforms all the univariate filter feature selection techniques and the evolutionary algorithms. It has low computational complexity in comparison to traditional wrapper-based evolutionary methods. The performance of the proposed approach is superior in comparison to existing unimodal and multimodal automated depression detection models. The combination of spectral, temporal and spectro-temporal speech features gave the best result with the LDA classifier. The performance achieved with the proposed approach, in terms of F1-score for the depressed class and the non-depressed class and error is 0.846, 0.932 and 0.094 respectively. Statistical tests demonstrate that the acoustic features selected using the proposed approach are non-redundant and discriminatory. Statistical tests also establish that the performance of the proposed approach is significantly better than that of the traditional wrapper-based evolutionary methods.
Collapse
Affiliation(s)
| | - Swati Rathi
- School of Computer and Systems Sciences, Jawaharlal Nehru University, Delhi, India.
| | - R K Agrawal
- School of Computer and Systems Sciences, Jawaharlal Nehru University, Delhi, India.
| |
Collapse
|
15
|
Li M, Liu Y, Liu Y, Pu C, Yin R, Zeng Z, Deng L, Wang X. Resting-state EEG-based convolutional neural network for the diagnosis of depression and its severity. Front Physiol 2022; 13:956254. [PMID: 36299253 PMCID: PMC9589234 DOI: 10.3389/fphys.2022.956254] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2022] [Accepted: 09/20/2022] [Indexed: 11/13/2022] Open
Abstract
Purpose: The study aimed to assess the value of the resting-state electroencephalogram (EEG)-based convolutional neural network (CNN) method for the diagnosis of depression and its severity in order to better serve depressed patients and at-risk populations. Methods: In this study, we used the resting state EEG-based CNN to identify depression and evaluated its severity. The EEG data were collected from depressed patients and healthy people using the Nihon Kohden EEG-1200 system. Analytical processing of resting-state EEG data was performed using Python and MATLAB software applications. The questionnaire included the Self-Rating Anxiety Scale (SAS), Self-Rating Depression Scale (SDS), Symptom Check-List-90 (SCL-90), and the Eysenck Personality Questionnaire (EPQ). Results: A total of 82 subjects were included in this study, with 41 in the depression group and 41 in the healthy control group. The area under the curve (AUC) of the resting-state EEG-based CNN in depression diagnosis was 0.74 (95%CI: 0.70–0.77) with an accuracy of 66.40%. In the depression group, the SDS, SAS, SCL-90 subscales, and N scores were significantly higher in the major depression group than those in the non-major depression group (p < 0.05). The AUC of the model in depression severity was 0.70 (95%CI: 0.65–0.75) with an accuracy of 66.93%. Correlation analysis revealed that major depression AI scores were significantly correlated with SAS scores (r = 0.508, p = 0.003) and SDS scores (r = 0.765, p < 0.001). Conclusion: Our model can accurately identify the depression-specific EEG signal in terms of depression diagnosis and severity identification. It would eventually provide new strategies for early diagnosis of depression and its severity.
Collapse
Affiliation(s)
- Mengqian Li
- Department of Psychosomatic Medicine, The First Affiliated Hospital of Nanchang University, Nanchang, China
| | - Yuan Liu
- Department of Psychosomatic Medicine, The First Affiliated Hospital of Nanchang University, Nanchang, China
| | - Yan Liu
- Second Clinical Medical College, Nanchang University, Nanchang, China
| | - Changqin Pu
- Queen Mary College, Nanchang University, Nanchang, China
| | - Ruocheng Yin
- Queen Mary College, Nanchang University, Nanchang, China
| | - Ziqiang Zeng
- Jiangxi Provincial Key Laboratory of Preventive Medicine, School of Public Health, Nanchang University, Nanchang, China
- School of Public Health, Nanchang University, Nanchang, China
| | - Libin Deng
- Jiangxi Provincial Key Laboratory of Preventive Medicine, School of Public Health, Nanchang University, Nanchang, China
- *Correspondence: Libin Deng, ; Xing Wang,
| | - Xing Wang
- School of Life Sciences, Nanchang University, Nanchang, China
- Clinical Medical Experiment Center, Nanchang University, Nanchang, China
- *Correspondence: Libin Deng, ; Xing Wang,
| |
Collapse
|
16
|
Gao X, Ma K, Yang H, Wang K, Fu B, Zhu Y, She X, Cui B. A rapid, non-invasive method for fatigue detection based on voice information. Front Cell Dev Biol 2022; 10:994001. [PMID: 36176279 PMCID: PMC9513181 DOI: 10.3389/fcell.2022.994001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2022] [Accepted: 08/24/2022] [Indexed: 11/19/2022] Open
Abstract
Fatigue results from a series of physiological and psychological changes due to continuous energy consumption. It can affect the physiological states of operators, thereby reducing their labor capacity. Fatigue can also reduce efficiency and, in serious cases, cause severe accidents. In addition, it can trigger pathological-related changes. By establishing appropriate methods to closely monitor the fatigue status of personnel and relieve the fatigue on time, operation-related injuries can be reduced. Existing fatigue detection methods mostly include subjective methods, such as fatigue scales, or those involving the use of professional instruments, which are more demanding for operators and cannot detect fatigue levels in real time. Speech contains information that can be used as acoustic biomarkers to monitor physiological and psychological statuses. In this study, we constructed a fatigue model based on the method of sleep deprivation by collecting various physiological indexes, such as P300 and glucocorticoid level in saliva, as well as fatigue questionnaires filled by 15 participants under different fatigue procedures and graded the fatigue levels accordingly. We then extracted the speech features at different instances and constructed a model to match the speech features and the degree of fatigue using a machine learning algorithm. Thus, we established a method to rapidly judge the degree of fatigue based on speech. The accuracy of the judgment based on unitary voice could reach 94%, whereas that based on long speech could reach 81%. Our fatigue detection method based on acoustic information can easily and rapidly determine the fatigue levels of the participants. This method can operate in real time and is non-invasive and efficient. Moreover, it can be combined with the advantages of information technology and big data to expand its applicability.
Collapse
Affiliation(s)
| | | | | | | | | | | | | | - Bo Cui
- *Correspondence: Xiaojun She, ; Bo Cui,
| |
Collapse
|
17
|
Online social network individual depression detection using a multitask heterogenous modality fusion approach. Inf Sci (N Y) 2022. [DOI: 10.1016/j.ins.2022.07.109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
18
|
Moukaddam N, Sano A, Salas R, Hammal Z, Sabharwal A. Turning data into better mental health: Past, present, and future. Front Digit Health 2022; 4:916810. [PMID: 36060543 PMCID: PMC9428351 DOI: 10.3389/fdgth.2022.916810] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2022] [Accepted: 07/21/2022] [Indexed: 11/13/2022] Open
Abstract
In this mini-review, we discuss the fundamentals of using technology in mental health diagnosis and tracking. We highlight those principles using two clinical concepts: (1) cravings and relapse in the context of addictive disorders and (2) anhedonia in the context of depression. This manuscript is useful for both clinicians wanting to understand the scope of technology use in psychiatry and for computer scientists and engineers wishing to assess psychiatric frameworks useful for diagnosis and treatment. The increase in smartphone ownership and internet connectivity, as well as the accelerated development of wearable devices, have made the observation and analysis of human behavior patterns possible. This has, in turn, paved the way to understand mental health conditions better. These technologies have immense potential in facilitating the diagnosis and tracking of mental health conditions; they also allow the implementation of existing behavioral treatments in new contexts (e.g., remotely, online, and in rural/underserved areas), and the possibility to develop new treatments based on new understanding of behavior patterns. The path to understand how to best use technology in mental health includes the need to match interdisciplinary frameworks from engineering/computer sciences and psychiatry. Thus, we start our review by introducing bio-behavioral sensing, the types of information available, and what behavioral patterns they may reflect and be related to in psychiatric diagnostic frameworks. This information is linked to the use of functional imaging, highlighting how imaging modalities can be considered "ground truth" for mental health/psychiatric dimensions, given the heterogeneity of clinical presentations, and the difficulty of determining what symptom corresponds to what disease. We then discuss how mental health/psychiatric dimensions overlap, yet differ from, psychiatric diagnoses. Using two clinical examples, we highlight the potential agreement areas in assessment/management of anhedonia and cravings. These two dimensions were chosen because of their link to two very prevalent diseases worldwide: depression and addiction. Anhedonia is a core symptom of depression, which is one of the leading causes of disability worldwide. Cravings, the urge to use a substance or perform an action (e.g., shopping, internet), is the leading step before relapse. Lastly, through the manuscript, we discuss potential mental health dimensions.
Collapse
Affiliation(s)
- Nidal Moukaddam
- Department of Psychiatry, Baylor College of Medicine, Houston Texas, United States
| | - Akane Sano
- Department of Electrical and Computer Engineering, Rice University, Houston, Texas, United States
| | - Ramiro Salas
- Department of Psychiatry, Baylor College of Medicine, The Menninger Clinic, Michael E DeBakey VA Medical Center, Houston, Texas, United States
| | - Zakia Hammal
- The Robotics Institute Department in the School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, United States
| | - Ashutosh Sabharwal
- Department of Electrical and Computer Engineering, Rice University, Houston, Texas, United States
| |
Collapse
|
19
|
Han J, Jiang G, Ouyang G, Li X. A Multimodal Approach for Identifying Autism Spectrum Disorders in Children. IEEE Trans Neural Syst Rehabil Eng 2022; 30:2003-2011. [PMID: 35853070 DOI: 10.1109/tnsre.2022.3192431] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Identification of autism spectrum disorder (ASD) in children is challenging due to the complexity and heterogeneity of ASD. Currently, most existing methods mainly rely on a single modality with limited information and often cannot achieve satisfactory performance. To address this issue, this paper investigates from internal neurophysiological and external behavior perspectives simultaneously and proposes a new multimodal diagnosis framework for identifying ASD in children with fusion of electroencephalogram (EEG) and eye-tracking (ET) data. Specifically, we designed a two-step multimodal feature learning and fusion model based on a typical deep learning algorithm, stacked denoising autoencoder (SDAE). In the first step, two SDAE models are designed for feature learning for EEG and ET modality, respectively. Then, a third SDAE model in the second step is designed to perform multimodal fusion with learned EEG and ET features in a concatenated way. Our designed multimodal identification model can automatically capture correlations and complementarity from behavior modality and neurophysiological modality in a latent feature space, and generate informative feature representations with better discriminability and generalization for enhanced identification performance. We collected a multimodal dataset containing 40 ASD children and 50 typically developing (TD) children to evaluate our proposed method. Experimental results showed that our proposed method achieved superior performance compared with two unimodal methods and a simple feature-level fusion method, which has promising potential to provide an objective and accurate diagnosis to assist clinicians.
Collapse
|
20
|
Yu Y, Li W, Zhao Y, Ye J, Zheng Y, Liu X, Wang Q. Depression and Severity Detection Based on Body Kinematic Features: Using Kinect Recorded Skeleton Data of Simple Action. Front Neurol 2022; 13:905917. [PMID: 35847201 PMCID: PMC9279697 DOI: 10.3389/fneur.2022.905917] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2022] [Accepted: 05/18/2022] [Indexed: 12/02/2022] Open
Abstract
Relative limb movement is an important feature in assessing depression. In this study, we looked into whether a skeleton-mimetic task using natural stimuli may help people recognize depression. We innovatively used Kinect V2 to collect participant data. Sequential skeletal data was directly extracted from the original Kinect-3D and tetrad coordinates of the participant's 25 body joints. Two constructed skeletal datasets of whole-body joints (including binary classification and multi classification) were input into the proposed model for depression recognition after data preparation. We improved the temporal convolution network (TCN), creating novel spatial attention dilated TCN (SATCN) network that included a hierarchy of temporal convolution groups with different dilated convolution scales to capture important skeletal features and a spatial attention block for final result prediction. The depression and non-depression groups can be classified automatically with a maximum accuracy of 75.8% in the binary classification task, and 64.3% accuracy in the multi classification dataset to recognize more fine-grained identification of depression severity, according to experimental results. Our experiments and methods based on Kinect V2 can not only identify and screen depression patients but also effectively observe the recovery level of depression patients during the recovery process. For example, in the change from severe depression to moderate or mild depression multi classification dataset.
Collapse
Affiliation(s)
- Yanhong Yu
- College of Traditional Chinese Medicine, Shandong University of Traditional Chinese Medicine, Jinan, China
| | - Wentao Li
- School of Computer Science and Technology, Qilu University of Technology, Jinan, China
| | - Yue Zhao
- The First Affiliated Hospital of Shandong First Medical University & Shandong Provincial Qianfoshan Hospital, Jinan, China
| | - Jiayu Ye
- School of Computer Science and Technology, Qilu University of Technology, Jinan, China
- Jiayu Ye
| | - Yunshao Zheng
- Shandong Mental Health Center, Shandong University, Jinan, China
| | - Xinxin Liu
- College of Traditional Chinese Medicine, Shandong University of Traditional Chinese Medicine, Jinan, China
| | - Qingxiang Wang
- School of Computer Science and Technology, Qilu University of Technology, Jinan, China
- *Correspondence: Qingxiang Wang
| |
Collapse
|
21
|
Wu P, Wang R, Lin H, Zhang F, Tu J, Sun M. Automatic depression recognition by intelligent speech signal processing: A systematic survey. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY 2022. [DOI: 10.1049/cit2.12113] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022] Open
Affiliation(s)
- Pingping Wu
- Jiangsu Key Laboratory of Public Project Audit, School of Engineering Audit Nanjing Audit University Nanjing China
| | - Ruihao Wang
- School of Information Engineering Nanjing Audit University Nanjing China
| | - Han Lin
- Jiangsu Key Laboratory of Public Project Audit, School of Engineering Audit Nanjing Audit University Nanjing China
| | - Fanlong Zhang
- School of Information Engineering Nanjing Audit University Nanjing China
| | - Juan Tu
- Key Laboratory of Modern Acoustics (MOE), School of Physics Nanjing University Nanjing China
| | - Miao Sun
- Faculty of Electrical Engineering, Mathematics & Computer Science Delft University of Technology Delft The Netherlands
| |
Collapse
|
22
|
Rong Q, Ding S, Yue Z, Wang Y, Wang L, Zheng X, Li Y. Non-Contact Negative Mood State Detection Using Reliability-Focused Multi-Modal Fusion Model. IEEE J Biomed Health Inform 2022; 26:4691-4701. [PMID: 35696474 DOI: 10.1109/jbhi.2022.3182357] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Negative mood states include tension, depression, anger, fatigue, and confusion, which represent the weak internal emotions of a human. Negative mood states exert adverse impact on individuals' ability to make rational decisions, which entails the practicable method of negative mood state detection. The most commonly used negative mood state detection methods are based on the psychological scale, which requires additional work and brings inconvenience to the subject in the application scenarios. To overcome this challenge, this paper proposes a novel non-contact negative mood state detection method according to the knowledge of affective computing. The POMS-net model is used to extract temporal-spatial features from visible and infrared thermal videos, and the negative mood state detection is realized using data reliability-focused multi-modal fusion. The proposed method is verified using the HDT-BR dataset collected in the aerospace medicine experiment "Earth-Star II" and the VIRI public dataset. The experimental results on the datasets verify that our method outperforms the comparison methods.
Collapse
|
23
|
Schultebraucks K, Yadav V, Shalev AY, Bonanno GA, Galatzer-Levy IR. Deep learning-based classification of posttraumatic stress disorder and depression following trauma utilizing visual and auditory markers of arousal and mood. Psychol Med 2022; 52:957-967. [PMID: 32744201 DOI: 10.1017/s0033291720002718] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/28/2022]
Abstract
BACKGROUND Visual and auditory signs of patient functioning have long been used for clinical diagnosis, treatment selection, and prognosis. Direct measurement and quantification of these signals can aim to improve the consistency, sensitivity, and scalability of clinical assessment. Currently, we investigate if machine learning-based computer vision (CV), semantic, and acoustic analysis can capture clinical features from free speech responses to a brief interview 1 month post-trauma that accurately classify major depressive disorder (MDD) and posttraumatic stress disorder (PTSD). METHODS N = 81 patients admitted to an emergency department (ED) of a Level-1 Trauma Unit following a life-threatening traumatic event participated in an open-ended qualitative interview with a para-professional about their experience 1 month following admission. A deep neural network was utilized to extract facial features of emotion and their intensity, movement parameters, speech prosody, and natural language content. These features were utilized as inputs to classify PTSD and MDD cross-sectionally. RESULTS Both video- and audio-based markers contributed to good discriminatory classification accuracy. The algorithm discriminates PTSD status at 1 month after ED admission with an AUC of 0.90 (weighted average precision = 0.83, recall = 0.84, and f1-score = 0.83) as well as depression status at 1 month after ED admission with an AUC of 0.86 (weighted average precision = 0.83, recall = 0.82, and f1-score = 0.82). CONCLUSIONS Direct clinical observation during post-trauma free speech using deep learning identifies digital markers that can be utilized to classify MDD and PTSD status.
Collapse
Affiliation(s)
- Katharina Schultebraucks
- Department of Emergency Medicine, Vagelos School of Physicians and Surgeons, Columbia University Irving Medical Center, New York, New York, USA
- Department of Psychiatry, New York University Grossman School of Medicine, New York, New York, USA
- Data Science Institute, Columbia University, New York, New York, USA
| | | | - Arieh Y Shalev
- Department of Psychiatry, New York University Grossman School of Medicine, New York, New York, USA
| | - George A Bonanno
- Department of Counseling and Clinical Psychology, Teachers College, Columbia University, New York, New York, USA
| | - Isaac R Galatzer-Levy
- Department of Psychiatry, New York University Grossman School of Medicine, New York, New York, USA
- AiCure, New York, New York, USA
| |
Collapse
|
24
|
Birnbaum ML, Abrami A, Heisig S, Ali A, Arenare E, Agurto C, Lu N, Kane JM, Cecchi G. Acoustic and Facial Features From Clinical Interviews for Machine Learning-Based Psychiatric Diagnosis: Algorithm Development. JMIR Ment Health 2022; 9:e24699. [PMID: 35072648 PMCID: PMC8822433 DOI: 10.2196/24699] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/01/2020] [Revised: 04/29/2021] [Accepted: 12/01/2021] [Indexed: 01/26/2023] Open
Abstract
BACKGROUND In contrast to all other areas of medicine, psychiatry is still nearly entirely reliant on subjective assessments such as patient self-report and clinical observation. The lack of objective information on which to base clinical decisions can contribute to reduced quality of care. Behavioral health clinicians need objective and reliable patient data to support effective targeted interventions. OBJECTIVE We aimed to investigate whether reliable inferences-psychiatric signs, symptoms, and diagnoses-can be extracted from audiovisual patterns in recorded evaluation interviews of participants with schizophrenia spectrum disorders and bipolar disorder. METHODS We obtained audiovisual data from 89 participants (mean age 25.3 years; male: 48/89, 53.9%; female: 41/89, 46.1%): individuals with schizophrenia spectrum disorders (n=41), individuals with bipolar disorder (n=21), and healthy volunteers (n=27). We developed machine learning models based on acoustic and facial movement features extracted from participant interviews to predict diagnoses and detect clinician-coded neuropsychiatric symptoms, and we assessed model performance using area under the receiver operating characteristic curve (AUROC) in 5-fold cross-validation. RESULTS The model successfully differentiated between schizophrenia spectrum disorders and bipolar disorder (AUROC 0.73) when aggregating face and voice features. Facial action units including cheek-raising muscle (AUROC 0.64) and chin-raising muscle (AUROC 0.74) provided the strongest signal for men. Vocal features, such as energy in the frequency band 1 to 4 kHz (AUROC 0.80) and spectral harmonicity (AUROC 0.78), provided the strongest signal for women. Lip corner-pulling muscle signal discriminated between diagnoses for both men (AUROC 0.61) and women (AUROC 0.62). Several psychiatric signs and symptoms were successfully inferred: blunted affect (AUROC 0.81), avolition (AUROC 0.72), lack of vocal inflection (AUROC 0.71), asociality (AUROC 0.63), and worthlessness (AUROC 0.61). CONCLUSIONS This study represents advancement in efforts to capitalize on digital data to improve diagnostic assessment and supports the development of a new generation of innovative clinical tools by employing acoustic and facial data analysis.
Collapse
Affiliation(s)
- Michael L Birnbaum
- Department of Psychiatry, The Zucker Hillside Hospital, Northwell Health, Glen Oaks, NY, United States.,The Feinstein Institute for Medical Research, Northwell Health, Manhasset, NY, United States.,The Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, NY, United States
| | - Avner Abrami
- Computational Biology Center, IBM Research, Yorktown Heights, NY, United States
| | - Stephen Heisig
- Icahn School of Medicine at Mount Sinai, New York City, NY, United States
| | - Asra Ali
- Department of Psychiatry, The Zucker Hillside Hospital, Northwell Health, Glen Oaks, NY, United States.,The Feinstein Institute for Medical Research, Northwell Health, Manhasset, NY, United States
| | - Elizabeth Arenare
- Department of Psychiatry, The Zucker Hillside Hospital, Northwell Health, Glen Oaks, NY, United States.,The Feinstein Institute for Medical Research, Northwell Health, Manhasset, NY, United States
| | - Carla Agurto
- Computational Biology Center, IBM Research, Yorktown Heights, NY, United States
| | - Nathaniel Lu
- Department of Psychiatry, The Zucker Hillside Hospital, Northwell Health, Glen Oaks, NY, United States.,The Feinstein Institute for Medical Research, Northwell Health, Manhasset, NY, United States
| | - John M Kane
- Department of Psychiatry, The Zucker Hillside Hospital, Northwell Health, Glen Oaks, NY, United States.,The Feinstein Institute for Medical Research, Northwell Health, Manhasset, NY, United States.,The Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, NY, United States
| | - Guillermo Cecchi
- Computational Biology Center, IBM Research, Yorktown Heights, NY, United States
| |
Collapse
|
25
|
Habib M, Wang Z, Qiu S, Zhao H, Murthy AS. Machine Learning Based Healthcare System for Investigating the Association Between Depression and Quality of Life. IEEE J Biomed Health Inform 2022; 26:2008-2019. [PMID: 34986108 DOI: 10.1109/jbhi.2022.3140433] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
New technological innovations are changing the future of healthcare system. Identification of factors that are responsible for causing depression may lead to new experiments and treatments. Because depression as a disease is becoming a leading community health concern worldwide. Using machine learning techniques this article presents a complete methodological framework to process and explore the heterogenous data and to better understand the association between factors related to quality of life and depression. Subsequently, the experimental study is mainly divided into two parts. In the first part, a data consolidation process is presented. The relationship of data is formed and to uniquely identify each relation in data the concept of the Secure Hash Algorithm is adopted. Hashing is used to locate and index the actual items in the data because it is easier to process short hash values instead of longer strings. The second part proposed a model using both unsupervised and supervised machine learning techniques. The consolidation approach helped in providing a base for formulation and validation of the research hypothesis. The Self organizing map provided 08 cluster solution and the classification problems were taken from the clustered data to further validate the performance of the posterior probability multi-class Support Vector Machine. The expectations of the importance sampling resulted in factors responsible for causing depression. The proposed model was adopted to improve the classification performance, and the result showed classification accuracy of 91.16%.
Collapse
|
26
|
Provenza NR, Sheth SA, Dastin-van Rijn EM, Mathura RK, Ding Y, Vogt GS, Avendano-Ortega M, Ramakrishnan N, Peled N, Gelin LFF, Xing D, Jeni LA, Ertugrul IO, Barrios-Anderson A, Matteson E, Wiese AD, Xu J, Viswanathan A, Harrison MT, Bijanki KR, Storch EA, Cohn JF, Goodman WK, Borton DA. Long-term ecological assessment of intracranial electrophysiology synchronized to behavioral markers in obsessive-compulsive disorder. Nat Med 2021; 27:2154-2164. [PMID: 34887577 PMCID: PMC8800455 DOI: 10.1038/s41591-021-01550-z] [Citation(s) in RCA: 32] [Impact Index Per Article: 10.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2021] [Accepted: 09/22/2021] [Indexed: 01/02/2023]
Abstract
Detection of neural signatures related to pathological behavioral states could enable adaptive deep brain stimulation (DBS), a potential strategy for improving efficacy of DBS for neurological and psychiatric disorders. This approach requires identifying neural biomarkers of relevant behavioral states, a task best performed in ecologically valid environments. Here, in human participants with obsessive-compulsive disorder (OCD) implanted with recording-capable DBS devices, we synchronized chronic ventral striatum local field potentials with relevant, disease-specific behaviors. We captured over 1,000 h of local field potentials in the clinic and at home during unstructured activity, as well as during DBS and exposure therapy. The wide range of symptom severity over which the data were captured allowed us to identify candidate neural biomarkers of OCD symptom intensity. This work demonstrates the feasibility and utility of capturing chronic intracranial electrophysiology during daily symptom fluctuations to enable neural biomarker identification, a prerequisite for future development of adaptive DBS for OCD and other psychiatric disorders.
Collapse
Affiliation(s)
- Nicole R Provenza
- Brown University School of Engineering, Providence, RI, USA
- Charles Stark Draper Laboratory, Cambridge, MA, USA
| | - Sameer A Sheth
- Department of Neurosurgery, Baylor College of Medicine, Houston, TX, USA
| | | | - Raissa K Mathura
- Department of Neurosurgery, Baylor College of Medicine, Houston, TX, USA
| | - Yaohan Ding
- Intelligent Systems Program, University of Pittsburgh, Pittsburgh, PA, USA
| | - Gregory S Vogt
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
| | - Michelle Avendano-Ortega
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
| | - Nithya Ramakrishnan
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
| | - Noam Peled
- MGH/HST Martinos Center for Biomedical Imaging, Charlestown, MA, USA
- Harvard Medical School, Cambridge, MA, USA
| | | | - David Xing
- Brown University School of Engineering, Providence, RI, USA
| | - Laszlo A Jeni
- Robotics Institute, Carnegie Mellon University, Pittsburgh, PA, USA
| | - Itir Onal Ertugrul
- Department of Cognitive Science and Artificial Intelligence, Tilburg University, Tilburg, the Netherlands
| | | | - Evan Matteson
- Brown University School of Engineering, Providence, RI, USA
| | - Andrew D Wiese
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
- Department of Psychology, University of Missouri-Kansas City, Kansas City, MO, USA
| | - Junqian Xu
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
- Department of Radiology, Baylor College of Medicine, Houston, TX, USA
| | - Ashwin Viswanathan
- Department of Neurosurgery, Baylor College of Medicine, Houston, TX, USA
| | | | - Kelly R Bijanki
- Department of Neurosurgery, Baylor College of Medicine, Houston, TX, USA
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
| | - Eric A Storch
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
| | - Jeffrey F Cohn
- Department of Psychology, University of Pittsburgh, Pittsburgh, PA, USA
| | - Wayne K Goodman
- Menninger Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, USA
| | - David A Borton
- Brown University School of Engineering, Providence, RI, USA.
- Carney Institute for Brain Science, Brown University, Providence, RI, USA.
- Center for Neurorestoration and Neurotechnology, Rehabilitation R&D Service, Department of Veterans Affairs, Providence, RI, USA.
| |
Collapse
|
27
|
Aydemir E, Tuncer T, Dogan S, Gururajan R, Acharya UR. Automated major depressive disorder detection using melamine pattern with EEG signals. APPL INTELL 2021. [DOI: 10.1007/s10489-021-02426-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
28
|
Ceccarelli F, Mahmoud M. Multimodal temporal machine learning for Bipolar Disorder and Depression Recognition. Pattern Anal Appl 2021. [DOI: 10.1007/s10044-021-01001-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
29
|
Chen X, Pan Z. A Convenient and Low-Cost Model of Depression Screening and Early Warning Based on Voice Data Using for Public Mental Health. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2021; 18:6441. [PMID: 34198659 PMCID: PMC8296267 DOI: 10.3390/ijerph18126441] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/20/2021] [Revised: 06/10/2021] [Accepted: 06/10/2021] [Indexed: 12/12/2022]
Abstract
Depression is a common mental health disease, which has great harm to public health. At present, the diagnosis of depression mainly depends on the interviews between doctors and patients, which is subjective, slow and expensive. Voice data are a kind of data that are easy to obtain and have the advantage of low cost. It has been proved that it can be used in the diagnosis of depression. The voice data used for modeling in this study adopted the authoritative public data set, which had passed the ethical review. The features of voice data were extracted by Python programming, and the voice features were stored in the format of CSV files. Through data processing, a big database, containing 1479 voice feature samples, was generated for modeling. Then, the decision tree screening model of depression was established by 10-fold cross validation and algorithm selection. The experiment achieved 83.4% prediction accuracy on voice data set. According to the prediction results of the model, the patients can be given early warning and intervention in time, so as to realize the health management of personal depression.
Collapse
Affiliation(s)
- Xin Chen
- School of Medicine, Hangzhou Normal University, Hangzhou 311121, China;
- Engineering Research Center of Mobile Health Management System, Ministry of Education, Hangzhou Normal University, Hangzhou 311121, China
- Institute of VR and Intelligent System, Hangzhou Normal University, Hangzhou 311121, China
| | - Zhigeng Pan
- School of Medicine, Hangzhou Normal University, Hangzhou 311121, China;
- Engineering Research Center of Mobile Health Management System, Ministry of Education, Hangzhou Normal University, Hangzhou 311121, China
- Institute of VR and Intelligent System, Hangzhou Normal University, Hangzhou 311121, China
| |
Collapse
|
30
|
Galatzer-Levy I, Abbas A, Ries A, Homan S, Sels L, Koesmahargyo V, Yadav V, Colla M, Scheerer H, Vetter S, Seifritz E, Scholz U, Kleim B. Validation of Visual and Auditory Digital Markers of Suicidality in Acutely Suicidal Psychiatric Inpatients: Proof-of-Concept Study. J Med Internet Res 2021; 23:e25199. [PMID: 34081022 PMCID: PMC8212625 DOI: 10.2196/25199] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2020] [Revised: 12/15/2020] [Accepted: 03/16/2021] [Indexed: 11/13/2022] Open
Abstract
BACKGROUND Multiple symptoms of suicide risk have been assessed based on visual and auditory information, including flattened affect, reduced movement, and slowed speech. Objective quantification of such symptomatology from novel data sources can increase the sensitivity, scalability, and timeliness of suicide risk assessment. OBJECTIVE We aimed to examine measurements extracted from video interviews using open-source deep learning algorithms to quantify facial, vocal, and movement behaviors in relation to suicide risk severity in recently admitted patients following a suicide attempt. METHODS We utilized video to quantify facial, vocal, and movement markers associated with mood, emotion, and motor functioning from a structured clinical conversation in 20 patients admitted to a psychiatric hospital following a suicide risk attempt. Measures were calculated using open-source deep learning algorithms for processing facial expressivity, head movement, and vocal characteristics. Derived digital measures of flattened affect, reduced movement, and slowed speech were compared to suicide risk with the Beck Scale for Suicide Ideation controlling for age and sex, using multiple linear regression. RESULTS Suicide severity was associated with multiple visual and auditory markers, including speech prevalence (β=-0.68, P=.02, r2=0.40), overall expressivity (β=-0.46, P=.10, r2=0.27), and head movement measured as head pitch variability (β=-1.24, P=.006, r2=0.48) and head yaw variability (β=-0.54, P=.06, r2=0.32). CONCLUSIONS Digital measurements of facial affect, movement, and speech prevalence demonstrated strong effect sizes and linear associations with the severity of suicidal ideation.
Collapse
Affiliation(s)
- Isaac Galatzer-Levy
- Research and Development, AiCure, New York, NY, United States
- Psychiatry, New York University School of Medicine, New York, NY, United States
| | - Anzar Abbas
- Research and Development, AiCure, New York, NY, United States
| | - Anja Ries
- Department of Psychology, University of Zurich, Zurich, Switzerland
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
- Neuroscience Centre Zurich, University of Zurich, Zurich, Switzerland
| | - Stephanie Homan
- Department of Psychology, University of Zurich, Zurich, Switzerland
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
- Neuroscience Centre Zurich, University of Zurich, Zurich, Switzerland
| | - Laura Sels
- Department of Psychology, University of Zurich, Zurich, Switzerland
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
| | | | - Vijay Yadav
- Research and Development, AiCure, New York, NY, United States
| | - Michael Colla
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
| | - Hanne Scheerer
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
| | - Stefan Vetter
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
| | - Erich Seifritz
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
- Neuroscience Centre Zurich, University of Zurich, Zurich, Switzerland
| | - Urte Scholz
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
| | - Birgit Kleim
- Department of Psychology, University of Zurich, Zurich, Switzerland
- Department of Psychiatry, Psychotherapy and Psychosomatics, University of Zurich, Zurich, Switzerland
- Neuroscience Centre Zurich, University of Zurich, Zurich, Switzerland
| |
Collapse
|
31
|
Abbas A, Sauder C, Yadav V, Koesmahargyo V, Aghjayan A, Marecki S, Evans M, Galatzer-Levy IR. Remote Digital Measurement of Facial and Vocal Markers of Major Depressive Disorder Severity and Treatment Response: A Pilot Study. Front Digit Health 2021; 3:610006. [PMID: 34713091 PMCID: PMC8521884 DOI: 10.3389/fdgth.2021.610006] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2020] [Accepted: 02/19/2021] [Indexed: 12/21/2022] Open
Abstract
Objectives: Multiple machine learning-based visual and auditory digital markers have demonstrated associations between major depressive disorder (MDD) status and severity. The current study examines if such measurements can quantify response to antidepressant treatment (ADT) with selective serotonin reuptake inhibitors (SSRIs) and serotonin-norepinephrine uptake inhibitors (SNRIs). Methods: Visual and auditory markers were acquired through an automated smartphone task that measures facial, vocal, and head movement characteristics across 4 weeks of treatment (with time points at baseline, 2 weeks, and 4 weeks) on ADT (n = 18). MDD diagnosis was confirmed using the Mini-International Neuropsychiatric Interview (MINI), and the Montgomery-Åsberg Depression Rating Scale (MADRS) was collected concordantly to assess changes in MDD severity. Results: Patient responses to ADT demonstrated clinically and statistically significant changes in the MADRS [F (2, 34) = 51.62, p < 0.0001]. Additionally, patients demonstrated significant increases in multiple digital markers including facial expressivity, head movement, and amount of speech. Finally, patients demonstrated significantly decreased frequency of fear and anger facial expressions. Conclusion: Digital markers associated with MDD demonstrate validity as measures of treatment response.
Collapse
Affiliation(s)
| | - Colin Sauder
- Adams Clinical, Watertown, MA, United States
- Karuna Therapeutics, Boston, MA, United States
| | | | | | | | | | | | - Isaac R. Galatzer-Levy
- AiCure, New York, NY, United States
- Psychiatry, New York University School of Medicine, New York, NY, United States
| |
Collapse
|
32
|
Abstract
In this chapter, I address the topic of tinnitus in the context of the patient's trajectory of care, with special attention to psychological comorbidities. Although most patients will cope with tinnitus and need no more than information and reassurance from professionals, a proportion of patients will need more supportive management. Assessment of psychological comorbidities is important to determine how urgent they should be seen in the clinic and their specific needs. The most frequent complaints are stress, depression, and anxiety. Although the direction of this relationship is still unclear (are comorbidities at the origin of tinnitus or are they a consequence of it), it is evident that the more serious comorbidities are at the onset of tinnitus, the worse the prognosis. Therefore, an assessment at the initial visit in the clinic is of utmost importance. There are valid and reliable psychometric tools to quickly draw a portrait of the psychological state of patients that can be used by audiologists, psychologists, or doctors. Therapeutic avenues can then be discussed with the patients to ensure them the best support possible.
Collapse
|
33
|
Toisoul A, Kossaifi J, Bulat A, Tzimiropoulos G, Pantic M. Estimation of continuous valence and arousal levels from faces in naturalistic conditions. NAT MACH INTELL 2021. [DOI: 10.1038/s42256-020-00280-0] [Citation(s) in RCA: 24] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
34
|
Smith E, Storch EA, Vahia I, Wong STC, Lavretsky H, Cummings JL, Eyre HA. Affective Computing for Late-Life Mood and Cognitive Disorders. Front Psychiatry 2021; 12:782183. [PMID: 35002802 PMCID: PMC8732874 DOI: 10.3389/fpsyt.2021.782183] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/23/2021] [Accepted: 11/29/2021] [Indexed: 12/20/2022] Open
Abstract
Affective computing (also referred to as artificial emotion intelligence or emotion AI) is the study and development of systems and devices that can recognize, interpret, process, and simulate emotion or other affective phenomena. With the rapid growth in the aging population around the world, affective computing has immense potential to benefit the treatment and care of late-life mood and cognitive disorders. For late-life depression, affective computing ranging from vocal biomarkers to facial expressions to social media behavioral analysis can be used to address inadequacies of current screening and diagnostic approaches, mitigate loneliness and isolation, provide more personalized treatment approaches, and detect risk of suicide. Similarly, for Alzheimer's disease, eye movement analysis, vocal biomarkers, and driving and behavior can provide objective biomarkers for early identification and monitoring, allow more comprehensive understanding of daily life and disease fluctuations, and facilitate an understanding of behavioral and psychological symptoms such as agitation. To optimize the utility of affective computing while mitigating potential risks and ensure responsible development, ethical development of affective computing applications for late-life mood and cognitive disorders is needed.
Collapse
Affiliation(s)
- Erin Smith
- The PRODEO Institute, San Francisco, CA, United States.,Organisation for Economic Co-operation and Development (OECD), Paris, France.,Department of Neurology & Neurological Sciences, Stanford University, Stanford, CA, United States.,Global Brain Health Institute, University of California, San Francisco, San Francisco, CA, United States.,Global Brain Health Institute, Trinity College Dublin, Dublin, Ireland
| | - Eric A Storch
- Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, United States
| | - Ipsit Vahia
- Division of Geriatric Psychiatry, McLean Hospital, Boston, MA, United States.,Division of Geriatric Psychiatry, Harvard Medical School, Boston, MA, United States
| | - Stephen T C Wong
- Systems Medicine and Biomedical Engineering Houston Methodist, Houston, TX, United States
| | - Helen Lavretsky
- Semel Institute for Neuroscience and Human Behavior, University of California, Los Angeles, Los Angeles, CA, United States
| | - Jeffrey L Cummings
- Chambers-Grundy Center for Transformative Neuroscience, Department of Brain Health, School of Integrated Health Sciences, University of Nevada, Las Vegas (UNLV), Las Vegas, NV, United States
| | - Harris A Eyre
- The PRODEO Institute, San Francisco, CA, United States.,Organisation for Economic Co-operation and Development (OECD), Paris, France.,Global Brain Health Institute, University of California, San Francisco, San Francisco, CA, United States.,Global Brain Health Institute, Trinity College Dublin, Dublin, Ireland.,Department of Psychiatry and Behavioral Sciences, Baylor College of Medicine, Houston, TX, United States.,IMPACT, The Institute for Mental and Physical Health and Clinical Translation, Deakin University, Geelong, VIC, Australia
| |
Collapse
|
35
|
Ding Y, Ertugrul IO, Darzi A, Provenza N, Jeni LA, Borton D, Goodman W, Cohn J. Automated Detection of Enhanced DBS Device Settings. COMPANION PUBLICATION OF THE 2020 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION 2020; 2020:354-356. [PMID: 33937916 DOI: 10.1145/3395035.3425354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
Continuous deep brain stimulation (DBS) of the ventral striatum (VS) is an effective treatment for severe, treatment-refractory obsessive-compulsive disorder (OCD). Optimal parameter settings are signaled by a mirth response of intense positive affect, which are subjectively identified by clinicians. Subjective judgments are idiosyncratic and difficult to standardize. To objectively measure mirth responses, we used Automatic Facial Affect Recognition (AFAR) in a series of longitudinal assessments of a patient treated with DBS. Pre- and post-adjustment DBS were compared using both statistical and machine learning approaches. Positive affect was significantly higher post-DBS adjustment. Using SVM and XGBoost, participant's pre- and post-adjustment appearances were differentiated with F1 of 0.76, which suggests feasibility of objective measurement of mirth response.
Collapse
Affiliation(s)
| | | | - Ali Darzi
- University of Pittsburgh, Pittsburgh, U.S.A
| | | | | | | | | | | |
Collapse
|
36
|
Hammal Z, Huang D, Bailly K, Chen L, Daoudi M. Face and Gesture Analysis for Health Informatics. PROCEEDINGS OF THE ... ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION. ICMI (CONFERENCE) 2020; 2020:874-875. [PMID: 33274351 PMCID: PMC7710162 DOI: 10.1145/3382507.3419747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
The goal of Face and Gesture Analysis for Health Informatics's workshop is to share and discuss the achievements as well as the challenges in using computer vision and machine learning for automatic human behavior analysis and modeling for clinical research and healthcare applications. The workshop aims to promote current research and support growth of multidisciplinary collaborations to advance this groundbreaking research. The meeting gathers scientists working in related areas of computer vision and machine learning, multi-modal signal processing and fusion, human centered computing, behavioral sensing, assistive technologies, and medical tutoring systems for healthcare applications and medicine.
Collapse
Affiliation(s)
- Zakia Hammal
- Carnegie Mellon University, Pittsburgh, Pennsylvania, USA
| | - Di Huang
- Beihang University, Beijing, China
| | | | | | | |
Collapse
|
37
|
Muszynski M, Zelazny J, Girard JM, Morency LP. Depression Severity Assessment for Adolescents at High Risk of Mental Disorders. PROCEEDINGS OF THE ... ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION. ICMI (CONFERENCE) 2020; 2020:70-78. [PMID: 33782675 PMCID: PMC8005296 DOI: 10.1145/3382507.3418859] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
Recent progress in artificial intelligence has led to the development of automatic behavioral marker recognition, such as facial and vocal expressions. Those automatic tools have enormous potential to support mental health assessment, clinical decision making, and treatment planning. In this paper, we investigate nonverbal behavioral markers of depression severity assessed during semi-structured medical interviews of adolescent patients. The main goal of our research is two-fold: studying a unique population of adolescents at high risk of mental disorders and differentiating mild depression from moderate or severe depression. We aim to explore computationally inferred facial and vocal behavioral responses elicited by three segments of the semi-structured medical interviews: Distress Assessment Questions, Ubiquitous Questions, and Concept Questions. Our experimental methodology reflects best practise used for analyzing small sample size and unbalanced datasets of unique patients. Our results show a very interesting trend with strongly discriminative behavioral markers from both acoustic and visual modalities. These promising results are likely due to the unique classification task (mild depression vs. moderate and severe depression) and three types of probing questions.
Collapse
|
38
|
Aafjes-van Doorn K, Kamsteeg C, Bate J, Aafjes M. A scoping review of machine learning in psychotherapy research. Psychother Res 2020; 31:92-116. [PMID: 32862761 DOI: 10.1080/10503307.2020.1808729] [Citation(s) in RCA: 62] [Impact Index Per Article: 15.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/23/2022] Open
Abstract
Machine learning (ML) offers robust statistical and probabilistic techniques that can help to make sense of large amounts of data. This scoping review paper aims to broadly explore the nature of research activity using ML in the context of psychological talk therapies, highlighting the scope of current methods and considerations for clinical practice and directions for future research. Using a systematic search methodology, fifty-one studies were identified. A narrative synthesis indicates two types of studies, those who developed and tested an ML model (k=44), and those who reported on the feasibility of a particular treatment tool that uses an ML algorithm (k=7). Most model development studies used supervised learning techniques to classify or predict labeled treatment process or outcome data, whereas others used unsupervised techniques to identify clusters in the unlabeled patient or treatment data. Overall, the current applications of ML in psychotherapy research demonstrated a range of possible benefits for indications of treatment process, adherence, therapist skills and treatment response prediction, as well as ways to accelerate research through automated behavioral or linguistic process coding. Given the novelty and potential of this research field, these proof-of-concept studies are encouraging, however, do not necessarily translate to improved clinical practice (yet).
Collapse
Affiliation(s)
| | | | - Jordan Bate
- Ferkauf Graduate School of Psychology, Yeshiva University, Bronx, NY, USA
| | | |
Collapse
|
39
|
Su MH, Wu CH, Huang KY, Yang TH. Cell-Coupled Long Short-Term Memory With L -Skip Fusion Mechanism for Mood Disorder Detection Through Elicited Audiovisual Features. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:124-135. [PMID: 30892247 DOI: 10.1109/tnnls.2019.2899884] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
In early stages, patients with bipolar disorder are often diagnosed as having unipolar depression in mood disorder diagnosis. Because the long-term monitoring is limited by the delayed detection of mood disorder, an accurate and one-time diagnosis is desirable to avoid delay in appropriate treatment due to misdiagnosis. In this paper, an elicitation-based approach is proposed for realizing a one-time diagnosis by using responses elicited from patients by having them watch six emotion-eliciting videos. After watching each video clip, the conversations, including patient facial expressions and speech responses, between the participant and the clinician conducting the interview were recorded. Next, the hierarchical spectral clustering algorithm was employed to adapt the facial expression and speech response features by using the extended Cohn-Kanade and eNTERFACE databases. A denoizing autoencoder was further applied to extract the bottleneck features of the adapted data. Then, the facial and speech bottleneck features were input into support vector machines to obtain speech emotion profiles (EPs) and the modulation spectrum (MS) of the facial action unit sequence for each elicited response. Finally, a cell-coupled long short-term memory (LSTM) network with an L -skip fusion mechanism was proposed to model the temporal information of all elicited responses and to loosely fuse the EPs and the MS for conducting mood disorder detection. The experimental results revealed that the cell-coupled LSTM with the L -skip fusion mechanism has promising advantages and efficacy for mood disorder detection.
Collapse
|
40
|
Bhatia S, Goecke R, Hammal Z, Cohn JF. Automated Measurement of Head Movement Synchrony during Dyadic Depression Severity Interviews. PROCEEDINGS OF THE ... INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION. IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION 2019; 2019. [PMID: 31745390 DOI: 10.1109/fg.2019.8756509] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
With few exceptions, most research in automated assessment of depression has considered only the patient's behavior to the exclusion of the therapist's behavior. We investigated the interpersonal coordination (synchrony) of head movement during patient-therapist clinical interviews. Our sample consisted of patients diagnosed with major depressive disorder. They were recorded in clinical interviews (Hamilton Rating Scale for Depression, HRSD) at 7-week intervals over a period of 21 weeks. For each session, patient and therapist 3D head movement was tracked from 2D videos. Head angles in the horizontal (pitch) and vertical (yaw) axes were used to measure head movement. Interpersonal coordination of head movement between patients and therapists was measured using windowed cross-correlation. Patterns of coordination in head movement were investigated using the peak picking algorithm. Changes in head movement coordination over the course of treatment were measured using a hierarchical linear model (HLM). The results indicated a strong effect for patient-therapist head movement synchrony. Within-dyad variability in head movement coordination was found to be higher than between-dyad variability, meaning that differences over time in a dyad were higher as compared to the differences between dyads. Head movement synchrony did not change over the course of treatment with change in depression severity. To the best of our knowledge, this study is the first attempt to analyze the mutual influence of patient-therapist head movement in relation to depression severity.
Collapse
Affiliation(s)
- Shalini Bhatia
- Human-Centred Technology Research Centre, University of Canberra, Canberra, Australia
| | - Roland Goecke
- Human-Centred Technology Research Centre, University of Canberra, Canberra, Australia
| | - Zakia Hammal
- Robotics Institute, Carnegie Mellon University, Pittsburgh, USA
| | - Jeffrey F Cohn
- Department of Psychology, University of Pittsburgh, Pittsburgh, USA
| |
Collapse
|
41
|
Wang J, Zhang L, Liu T, Pan W, Hu B, Zhu T. Acoustic differences between healthy and depressed people: a cross-situation study. BMC Psychiatry 2019; 19:300. [PMID: 31615470 PMCID: PMC6794822 DOI: 10.1186/s12888-019-2300-7] [Citation(s) in RCA: 27] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/06/2018] [Accepted: 09/20/2019] [Indexed: 11/29/2022] Open
Abstract
BACKGROUND Abnormalities in vocal expression during a depressed episode have frequently been reported in people with depression, but less is known about if these abnormalities only exist in special situations. In addition, the impacts of irrelevant demographic variables on voice were uncontrolled in previous studies. Therefore, this study compares the vocal differences between depressed and healthy people under various situations with irrelevant variables being regarded as covariates. METHODS To examine whether the vocal abnormalities in people with depression only exist in special situations, this study compared the vocal differences between healthy people and patients with unipolar depression in 12 situations (speech scenarios). Positive, negative and neutral voice expressions between depressed and healthy people were compared in four tasks. Multiple analysis of covariance (MANCOVA) was used for evaluating the main effects of variable group (depressed vs. healthy) on acoustic features. The significances of acoustic features were evaluated by both statistical significance and magnitude of effect size. RESULTS The results of multivariate analysis of covariance showed that significant differences between the two groups were observed in all 12 speech scenarios. Although significant acoustic features were not the same in different scenarios, we found that three acoustic features (loudness, MFCC5 and MFCC7) were consistently different between people with and without depression with large effect magnitude. CONCLUSIONS Vocal differences between depressed and healthy people exist in 12 scenarios. Acoustic features including loudness, MFCC5 and MFCC7 have potentials to be indicators for identifying depression via voice analysis. These findings support that depressed people's voices include both situation-specific and cross-situational patterns of acoustic features.
Collapse
Affiliation(s)
- Jingying Wang
- Institute of Psychology, Chinese Academy of Sciences, Beijing, China
| | - Lei Zhang
- Department of Computer Science, Virginia Tech, Blacksburg, VA USA
| | - Tianli Liu
- Institute of Population Research, Peking University, Beijing, China
| | - Wei Pan
- Institute of Psychology, Chinese Academy of Sciences, Beijing, China
| | - Bin Hu
- School of Information Science and Engineering, Lanzhou University, Lanzhou, Gansu Province China
| | - Tingshao Zhu
- Institute of Psychology, Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
42
|
Harati S, Crowell A, Huang Y, Mayberg H, Nemati S. Classifying Depression Severity in Recovery From Major Depressive Disorder via Dynamic Facial Features. IEEE J Biomed Health Inform 2019; 24:815-824. [PMID: 31352356 DOI: 10.1109/jbhi.2019.2930604] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Major depressive disorder is a common psychiatric illness. At present, there are no objective, non-verbal, automated markers that can reliably track treatment response. Here, we explore the use of video analysis of facial expressivity in a cohort of severely depressed patients before and after deep brain stimulation (DBS), an experimental treatment for depression. We introduced a set of variability measurements to obtain unsupervised features from muted video recordings, which were then leveraged to build predictive models to classify three levels of severity in the patients' recovery from depression. Multiscale entropy was utilized to estimate the variability in pixel intensity level at various time scales. A dynamic latent variable model was utilized to learn a low-dimensional representation of factors that describe the dynamic relationship between high-dimensional pixels in each video frame and over time. Finally, a novel elastic net ordinal regression model was trained to predict the severity of depression, as independently rated by standard rating scales. Our results suggest that unsupervised features extracted from these video recordings, when incorporated in an ordinal regression predictor, can discriminate different levels of depression severity during ongoing DBS treatment. Objective markers of patient response to treatment have the potential to standardize treatment protocols and enhance the design of future clinical trials.
Collapse
|
43
|
Rana R, Latif S, Gururajan R, Gray A, Mackenzie G, Humphris G, Dunn J. Automated screening for distress: A perspective for the future. Eur J Cancer Care (Engl) 2019; 28:e13033. [DOI: 10.1111/ecc.13033] [Citation(s) in RCA: 21] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2019] [Revised: 02/05/2019] [Accepted: 02/18/2019] [Indexed: 01/13/2023]
Affiliation(s)
- Rajib Rana
- University of Southern Queensland Springfield Queensland Australia
| | - Siddique Latif
- University of Southern Queensland Springfield Queensland Australia
| | - Raj Gururajan
- University of Southern Queensland Springfield Queensland Australia
| | - Anthony Gray
- University of Southern Queensland Springfield Queensland Australia
| | | | | | - Jeff Dunn
- University of Southern Queensland Springfield Queensland Australia
- Griffith University Brisbane Queensland Australia
- University of Technology Sydney Sydney New South Wales Australia
| |
Collapse
|
44
|
Cohn JF, Okun MS, Jeni LA, Ertugrul IO, Borton D, Malone D, Goodman WK. Automated Affect Detection in Deep Brain Stimulation for Obsessive-Compulsive Disorder: A Pilot Study. PROCEEDINGS OF THE ... ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION. ICMI (CONFERENCE) 2018; 2018:40-44. [PMID: 30511050 PMCID: PMC6271416 DOI: 10.1145/3242969.3243023] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
Abstract
Automated measurement of affective behavior in psychopathology has been limited primarily to screening and diagnosis. While useful, clinicians more often are concerned with whether patients are improving in response to treatment. Are symptoms abating, is affect becoming more positive, are unanticipated side effects emerging? When treatment includes neural implants, need for objective, repeatable biometrics tied to neurophysiology becomes especially pressing. We used automated face analysis to assess treatment response to deep brain stimulation (DBS) in two patients with intractable obsessive-compulsive disorder (OCD). One was assessed intraoperatively following implantation and activation of the DBS device. The other was assessed three months post-implantation. Both were assessed during DBS on and o conditions. Positive and negative valence were quantified using a CNN trained on normative data of 160 non-OCD participants. Thus, a secondary goal was domain transfer of the classifiers. In both contexts, DBS-on resulted in marked positive affect. In response to DBS-off, affect flattened in both contexts and alternated with increased negative affect in the outpatient setting. Mean AUC for domain transfer was 0.87. These findings suggest that parametric variation of DBS is strongly related to affective behavior and may introduce vulnerability for negative affect in the event that DBS is discontinued.
Collapse
|
45
|
Kacem A, Hammal Z, Daoudi M, Cohn J. Detecting Depression Severity by Interpretable Representations of Motion Dynamics. PROCEEDINGS OF THE ... INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION. IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION 2018; 2018:739-745. [PMID: 30271308 PMCID: PMC6157749 DOI: 10.1109/fg.2018.00116] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Recent breakthroughs in deep learning using automated measurement of face and head motion have made possible the first objective measurement of depression severity. While powerful, deep learning approaches lack interpretability. We developed an interpretable method of automatically measuring depression severity that uses barycentric coordinates of facial landmarks and a Lie-algebra based rotation matrix of 3D head motion. Using these representations, kinematic features are extracted, preprocessed, and encoded using Gaussian Mixture Models (GMM) and Fisher vector encoding. A multi-class SVM is used to classify the encoded facial and head movement dynamics into three levels of depression severity. The proposed approach was evaluated in adults with history of chronic depression. The method approached the classification accuracy of state-of-the-art deep learning while enabling clinically and theoretically relevant findings. The velocity and acceleration of facial movement strongly mapped onto depression severity symptoms consistent with clinical data and theory.
Collapse
Affiliation(s)
- Anis Kacem
- IMT Lille Douai, Univ. Lille, CNRS, UMR 9189 CRIStAL, F-59000 Lille, France
| | - Zakia Hammal
- Robotics Institute, Carnegie Mellon University, USA
| | - Mohamed Daoudi
- IMT Lille Douai, Univ. Lille, CNRS, UMR 9189 CRIStAL, F-59000 Lille, France
| | - Jeffrey Cohn
- Robotics Institute, Carnegie Mellon University, USA
- Department of Psychology, University of Pittsburgh, USA
| |
Collapse
|
46
|
Hammal Z, Cohn JF, Wallace ER, Heike CL, Birgfeld CB, Oster H, Speltz ML. Facial Expressiveness in Infants With and Without Craniofacial Microsomia: Preliminary Findings. Cleft Palate Craniofac J 2018; 55:711-720. [PMID: 29377723 PMCID: PMC5936082 DOI: 10.1177/1055665617753481] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
OBJECTIVE To compare facial expressiveness (FE) of infants with and without craniofacial macrosomia (cases and controls, respectively) and to compare phenotypic variation among cases in relation to FE. DESIGN Positive and negative affect was elicited in response to standardized emotion inductions, video recorded, and manually coded from video using the Facial Action Coding System for Infants and Young Children. SETTING Five craniofacial centers: Children's Hospital of Los Angeles, Children's Hospital of Philadelphia, Seattle Children's Hospital, University of Illinois-Chicago, and University of North Carolina-Chapel Hill. PARTICIPANTS Eighty ethnically diverse 12- to 14-month-old infants. MAIN OUTCOME MEASURES FE was measured on a frame-by-frame basis as the sum of 9 observed facial action units (AUs) representative of positive and negative affect. RESULTS FE differed between conditions intended to elicit positive and negative affect (95% confidence interval = 0.09-0.66, P = .01). FE failed to differ between cases and controls (ES = -0.16 to -0.02, P = .47 to .92). Among cases, those with and without mandibular hypoplasia showed similar levels of FE (ES = -0.38 to 0.54, P = .10 to .66). CONCLUSIONS FE varied between positive and negative affect, and cases and controls responded similarly. Null findings for case/control differences may be attributable to a lower than anticipated prevalence of nerve palsy among cases, the selection of AUs, or the use of manual coding. In future research, we will reexamine group differences using an automated, computer vision approach that can cover a broader range of facial movements and their dynamics.
Collapse
Affiliation(s)
- Zakia Hammal
- Robotics Institute, Carnegie Mellon University, Pittsburgh, PA, USA
| | - Jeffrey F. Cohn
- Robotics Institute, Carnegie Mellon University, Pittsburgh, PA, USA
- Department of Psychology, University of Pittsburgh, Pittsburgh, PA, USA
| | | | - Carrie L. Heike
- Seattle Children’s Research Institute, Seattle, WA, USA
- Seattle Children’s Hospital, Seattle, WA, USA
- University of Washington School of Medicine, Seattle, WA, USA
| | - Craig B. Birgfeld
- Seattle Children’s Research Institute, Seattle, WA, USA
- Seattle Children’s Hospital, Seattle, WA, USA
- University of Washington School of Medicine, Seattle, WA, USA
| | - Harriet Oster
- NYU School of Professional Studies, New York, NY, USA
| | - Matthew L. Speltz
- Seattle Children’s Research Institute, Seattle, WA, USA
- University of Washington School of Medicine, Seattle, WA, USA
| |
Collapse
|