Audio-Visual Weakly Supervised Approach for Apathy Detection in the Elderly

被引:0
作者
Sharma, Garima [1 ]
Joshi, Jyoti [1 ]
Zeghari, Radia [2 ]
Guerchouche, Rachid [3 ]
机构
[1] Monash Univ, Human Ctr Artificial Intelligence Grp, Clayton, Vic, Australia
[2] Univ Cote dAzur, CoBTeK Lab, Nice, France
[3] INRIA, Le Chesnay, France
来源
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2020年
关键词
Apathy detection; Emotion recognition; Multiple instance learning; Digital health; FACIAL EXPRESSIONS; DIAGNOSIS; DISORDERS;
D O I
10.1109/ijcnn48605.2020.9206829
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Apathy is manifested as lack of feelings or emotions in several neurological and psychological disorders. Hence, directly impairing the display of emotion through facial expressions and speech. Current practices of prediction of apathy heavily rely on clinical diagnosis, an expert interviewing a patient or reports from patients' family members. The dependence on an expert and the human bias in its examination results in under-diagnosis of condition. In this paper, a multimodal multi-instance learning based method is proposed for automatic apathy detection. There are several challenges present while automating the process. Some of which are - recognizing emotions in elderly people, correct identification of emotions in a conversation and identifying differences between emotional responses from apathetic and non apathetic cohorts. The proposed method uses the audio and visual information in a weakly supervised manner to learn the apathetic behaviour in order to address these challenges. Features from facial expressions, action units, facial landmarks and audio signals are extracted for training. The fusion of multiple modalities in a weakly supervised method achieves 75.71% accuracy for apathy detection in elderly people. The experiments show that multimodal fusion is able to leverage on the presence of complimentary information across different modalities.
引用
收藏
页数:7
相关论文
共 52 条
[1]  
Agarwal S., 2019, P IEEE C COMPUTER VI, P38, DOI [10.4108/eai.18-7-2019, DOI 10.4108/EAI.18-7-2019]
[2]  
Ahmed S., 2018, 2018 AAAI Spring Symposium Series
[3]  
Alugupally N., 2011, Pattern Recognition and Image Analysis, V21, P681
[4]  
[Anonymous], 2015, 2015 11 IEEE INT C W, DOI DOI 10.1109/FG.2015.7284869
[5]  
[Anonymous], 2014, C EMPIRICAL METHODS, DOI 10.3115/v1/d14-1179.
[6]  
[Anonymous], 2002, NeurIPS, DOI DOI 10.5555/2968618.2968690
[7]  
[Anonymous], 2012, Proceedings of the international joint conference on neural networks (IJCNN), DOI DOI 10.1109/2FIJCNN.2012.6252784
[8]   OpenFace 2.0: Facial Behavior Analysis Toolkit [J].
Baltrusaitis, Tadas ;
Zadeh, Amir ;
Lim, Yao Chong ;
Morency, Louis-Philippe .
PROCEEDINGS 2018 13TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2018), 2018, :59-66
[9]   Emotional Expressions Reconsidered: Challenges to Inferring Emotion From Human Facial Movements [J].
Barrett, Lisa Feldman ;
Adolphs, Ralph ;
Marsella, Stacy ;
Martinez, Aleix M. ;
Pollak, Seth D. .
PSYCHOLOGICAL SCIENCE IN THE PUBLIC INTEREST, 2019, 20 (01) :1-68
[10]   Multiple instance learning: A survey of problem characteristics and applications [J].
Carbonneau, Marc-Andre ;
Cheplygina, Veronika ;
Granger, Eric ;
Gagnon, Ghyslain .
PATTERN RECOGNITION, 2018, 77 :329-353