A modular machine learning tool for holistic and fine-grained behavioral analysis

被引:1
作者
Michelot, Bruno [1 ]
Corneyllie, Alexandra [1 ]
Thevenet, Marc [1 ]
Duffner, Stefan [2 ]
Perrin, Fabien [1 ]
机构
[1] Ctr Rech Neurosci Lyon, CAP Team, INSERM U1028, CNRS UMR 5292,UCBL,UJM, 95 Blvd Pinel, F-69675 Bron, France
[2] Univ Claude Bernard Lyon 1, Univ Lumiere Lyon 2, Ecole Cent Lyon,UMR 5205 CNRS,INSA Lyon, IMAGINE Team,Lab InfoRmat Image & Syst Informat, Lyon, France
关键词
Behavior; Computer vision; Machine learning; Explainability; EXPLAINABLE ARTIFICIAL-INTELLIGENCE; MOVEMENT; PERCEPTION; ATTENTION; TRACKING; SENSORS; BLINK;
D O I
10.3758/s13428-024-02511-3
中图分类号
B841 [心理学研究方法];
学科分类号
040201 ;
摘要
Artificial intelligence techniques offer promising avenues for exploring human body features from videos, yet no freely accessible tool has reliably provided holistic and fine-grained behavioral analyses to date. To address this, we developed a machine learning tool based on a two-level approach: a first lower-level processing using computer vision for extracting fine-grained and comprehensive behavioral features such as skeleton or facial points, gaze, and action units; a second level of machine learning classification coupled with explainability providing modularity, to determine which behavioral features are triggered by specific environments. To validate our tool, we filmed 16 participants across six conditions, varying according to the presence of a person ("Pers"), a sound ("Snd"), or silence ("Rest"), and according to emotional levels using self-referential ("Self") and control ("Ctrl") stimuli. We demonstrated the effectiveness of our approach by extracting and correcting behavior from videos using two computer vision software (OpenPose and OpenFace) and by training two algorithms (XGBoost and long short-term memory [LSTM]) to differentiate between experimental conditions. High classification rates were achieved for "Pers" conditions versus "Snd" or "Rest" (AUC = 0.8-0.9), with explainability revealing actions units and gaze as key features. Additionally, moderate classification rates were attained for "Snd" versus "Rest" (AUC = 0.7), attributed to action units, limbs and head points, as well as for "Self" versus "Ctrl" (AUC = 0.7-0.8), due to facial points. These findings were consistent with a more conventional hypothesis-driven approach. Overall, our study suggests that our tool is well suited for holistic and fine-grained behavioral analysis and offers modularity for extension into more complex naturalistic environments.
引用
收藏
页数:17
相关论文
共 92 条
[1]   Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) [J].
Adadi, Amina ;
Berrada, Mohammed .
IEEE ACCESS, 2018, 6 :52138-52160
[2]   Optuna: A Next-generation Hyperparameter Optimization Framework [J].
Akiba, Takuya ;
Sano, Shotaro ;
Yanase, Toshihiko ;
Ohta, Takeru ;
Koyama, Masanori .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :2623-2631
[3]   A Review on Computer Vision-Based Methods for Human Action Recognition [J].
Al-Faris, Mahmoud ;
Chiverton, John ;
Ndzi, David ;
Ahmed, Ahmed Isam .
JOURNAL OF IMAGING, 2020, 6 (06)
[4]   Person Perception, Meet People Perception: Exploring the Social Vision of Groups [J].
Alt, Nicholas P. ;
Phillips, L. Taylor .
PERSPECTIVES ON PSYCHOLOGICAL SCIENCE, 2022, 17 (03) :768-787
[5]   Movement and Emotional Facial Expressions during the Adult Attachment Interview: Interaction Effects of Attachment and Anxiety Disorder [J].
Altmann, Uwe ;
Friemann, Catharina ;
Frank, Theresa S. ;
Sittler, Mareike C. ;
Schoenherr, Desiree ;
Singh, Sashi ;
Schurig, Susan ;
Strauss, Bernhard ;
Petrowski, Katja .
PSYCHOPATHOLOGY, 2021, 54 (01) :47-58
[6]  
[Anonymous], 2022, JASP (Version 0.16.2) [Computer software]
[7]  
[Anonymous], 2009, International Journal of Information Technology and Knowledge Management
[8]  
Azhagusundari B., 2013, International Journal of Innovative Technology and Exploring Engineering, V2, P18, DOI DOI 10.1371/JOURNAL.PONE.0166017
[9]  
Badger M., 2020, 3D Bird Reconstruction: A Dataset, Model, and Shape Recovery from a Single View, V12363, P1, DOI [10.1007/978-3-030-58523-51, DOI 10.1007/978-3-030-58523-51]
[10]  
Baltrusaitis T, 2016, IEEE WINT CONF APPL