Facial Representation
Facial representation research focuses on developing robust and accurate methods for encoding and analyzing facial images and videos, aiming to improve applications like facial recognition, expression analysis, and avatar animation. Current research emphasizes learning effective facial representations using deep learning models, including convolutional neural networks, transformers, and autoencoders, often incorporating self-supervised learning and techniques like contrastive learning and masked image modeling to leverage large, unlabeled datasets. These advancements are significant for improving the accuracy and fairness of facial analysis systems across diverse applications, including healthcare, security, and human-computer interaction.
Papers
Universal Facial Encoding of Codec Avatars from VR Headsets
Shaojie Bai, Te-Li Wang, Chenghui Li, Akshay Venkatesh, Tomas Simon, Chen Cao, Gabriel Schwartz, Ryan Wrench, Jason Saragih, Yaser Sheikh, Shih-En Wei
Compound Expression Recognition via Multi Model Ensemble for the ABAW7 Challenge
Xuxiong Liu, Kang Shen, Jun Yao, Boyan Wang, Minrui Liu, Liuwei An, Zishun Cui, Weijie Feng, Xiao Sun