Co Speech 3D Gesture
Co-speech 3D gesture generation aims to create realistic and expressive 3D hand and body movements synchronized with spoken audio. Current research heavily utilizes diffusion models, often pre-trained on massive synthetic datasets to overcome data scarcity, and incorporates techniques like ControlNets and transformers to better align audio features with generated gestures, while also focusing on modeling emotional nuances and improving temporal coherence. This field is significant for advancing virtual avatar animation, human-computer interaction, and potentially improving communication technologies by creating more natural and engaging synthetic interactions.
Papers
May 27, 2024
April 30, 2024
June 20, 2023