Behavior Analysis in the Wild
Behavior analysis in the wild focuses on automatically recognizing human emotions and affective states from unconstrained real-world videos and audio. Current research heavily utilizes deep learning, particularly employing transformer networks, convolutional neural networks (CNNs), and recurrent neural networks (RNNs) like LSTMs, often in multi-task and multi-modal frameworks that integrate visual, audio, and even textual data. These advancements aim to improve the accuracy and robustness of emotion recognition systems, with significant implications for applications such as human-computer interaction, mental health monitoring, and personalized experiences.
Papers
Unimodal Multi-Task Fusion for Emotional Mimicry Intensity Prediction
Tobias Hallmen, Fabian Deuser, Norbert Oswald, Elisabeth André
HSEmotion Team at the 6th ABAW Competition: Facial Expressions, Valence-Arousal and Emotion Intensity Prediction
Andrey V. Savchenko
Zero-shot Compound Expression Recognition with Visual Language Model at the 6th ABAW Challenge
Jiahe Wang, Jiale Huang, Bingzhao Cai, Yifan Cao, Xin Yun, Shangfei Wang
Boosting Continuous Emotion Recognition with Self-Pretraining using Masked Autoencoders, Temporal Convolutional Networks, and Transformers
Weiwei Zhou, Jiada Lu, Chenkun Ling, Weifeng Wang, Shaowei Liu
Multi-modal Facial Action Unit Detection with Large Pre-trained Models for the 5th Competition on Affective Behavior Analysis in-the-wild
Yufeng Yin, Minh Tran, Di Chang, Xinrui Wang, Mohammad Soleymani
Spatial-temporal Transformer for Affective Behavior Analysis
Peng Zou, Rui Wang, Kehua Wen, Yasi Peng, Xiao Sun
A transformer-based approach to video frame-level prediction in Affective Behaviour Analysis In-the-wild
Dang-Khanh Nguyen, Ngoc-Huynh Ho, Sudarshan Pant, Hyung-Jeong Yang
A Dual Branch Network for Emotional Reaction Intensity Estimation
Jun Yu, Jichao Zhu, Wangyuan Zhu, Zhongpeng Cai, Guochen Xie, Renda Li, Gongpeng Zhao
Emotional Reaction Intensity Estimation Based on Multimodal Data
Shangfei Wang, Jiaqiang Wu, Feiyi Zheng, Xin Li, Xuewei Li, Suwen Wang, Yi Wu, Yanan Chang, Xiangyu Miao
Local Region Perception and Relationship Learning Combined with Feature Fusion for Facial Action Unit Detection
Jun Yu, Renda Li, Zhongpeng Cai, Gongpeng Zhao, Guochen Xie, Jichao Zhu, Wangyuan Zhu
Leveraging TCN and Transformer for effective visual-audio fusion in continuous emotion recognition
Weiwei Zhou, Jiada Lu, Zhaolong Xiong, Weifeng Wang
Coarse-to-Fine Cascaded Networks with Smooth Predicting for Video Facial Expression Recognition
Fanglei Xue, Zichang Tan, Yu Zhu, Zhongsong Ma, Guodong Guo
Random Forest Regression for continuous affect using Facial Action Units
Saurabh Hinduja, Shaun Canavan, Liza Jivnani, Sk Rahatul Jannat, V Sri Chakra Kumar