AAAI Publications, Twenty-Ninth AAAI Conference on Artificial Intelligence

Font Size: 
Tackling Mental Health by Integrating Unobtrusive Multimodal Sensing
Dawei Zhou, Jiebo Luo, Vincent M.B. Silenzio, Yun Zhou, Jile Hu, Glenn Currier, Henry Kautz

Last modified: 2015-02-16


Mental illness is becoming a major plague in modern societies and poses challenges to the capacity of current public health systems worldwide. With the widespread adoption of social media and mobile devices, and rapid advances in artificial intelligence, a unique opportunity arises for tackling mental health problems. In this study, we investigate how users’ online social activities and physiological signals detected through ubiquitous sensors can be utilized in realistic scenarios for monitoring their mental health states. First, we extract a suite of multimodal time-series signals using modern computer vision and signal processing techniques, from recruited participants while they are immersed in online social media that elicit emotions and emotion transitions. Next, we use machine learning techniques to build a model that establishes the connection between mental states and the extracted multimodal signals. Finally, we validate the effectiveness of our approach using two groups of recruited subjects.


mental health, affect signals, multimodal analysis, social media, computer vision, machine learning

Full Text: PDF