Vocal Burst
Vocal bursts, short non-speech vocalizations expressing emotion (e.g., laughter, cries), are a growing area of research in affective computing. Current efforts focus on developing robust machine learning models, often employing convolutional neural networks, self-supervised learning (like wav2vec2 and data2vec), and attention mechanisms, to classify burst types and predict emotional states (arousal, valence, and discrete emotion categories) from audio data. This research is significant because it advances our understanding of nonverbal emotional communication and has potential applications in areas like mental health monitoring, human-computer interaction, and social robotics. The availability of large, publicly accessible datasets like EmoGator is crucial for driving progress in this field.
Papers
Self-Relation Attention and Temporal Awareness for Emotion Recognition via Vocal Burst
Dang-Linh Trinh, Minh-Cong Vo, Guee-Sang Lee
Self-Supervised Attention Networks and Uncertainty Loss Weighting for Multi-Task Emotion Recognition on Vocal Bursts
Vincent Karas, Andreas Triantafyllopoulos, Meishu Song, Björn W. Schuller