Automated Screening for Bipolar Disorder from Audio/Visual Modalities.
MM '18: ACM Multimedia Conference Seoul Republic of Korea October, 2018(2018)
摘要
This paper addresses the Bipolar Disorder sub-challenge of the Audio/Visual Emotion recognition Challenge (AVEC) 2018, where the objective is to classify patients suffering from bipolar disorder into states of remission, hypo-mania, and mania, from audio-visual recordings of structured interviews. To this end, we propose 'turbulence features' to capture sudden, erratic changes in feature contours from audio and visual modalities, and demonstrate their efficacy for the task at hand. We introduce Fisher Vector encoding of ComParE low level descriptors (LLDs) and demonstrate that these features are viable for screening of bipolar disorder from speech. We also perform several experiments with standard feature sets from the OpenSmile toolkit as well as multi-modal fusion. The best result achieved on the test set is a UAR = 57.41%, which matches the best result published as the official baseline.
更多查看译文
关键词
Social Signal Processing, Affective Computing, Bipolar Disorder, Machine Learning, Automated Screening, AVEC 2018
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络