Spoken Language Understanding
Spoken Language Understanding (SLU) focuses on enabling computers to comprehend human speech, aiming to extract meaning and intent from spoken dialogue. Current research emphasizes improving the robustness and accuracy of SLU systems, particularly in handling noisy speech, low-resource languages, and out-of-distribution data, often employing large language models (LLMs) and contrastive learning techniques within various architectures like end-to-end models and hybrid approaches combining speech encoders with LLMs. Advances in SLU are crucial for enhancing human-computer interaction in applications such as virtual assistants, improving accessibility for diverse languages, and advancing the broader field of artificial intelligence.
Papers
A Dynamic Graph Interactive Framework with Label-Semantic Injection for Spoken Language Understanding
Zhihong Zhu, Weiyuan Xu, Xuxin Cheng, Tengtao Song, Yuexian Zou
Robust Unstructured Knowledge Access in Conversational Dialogue with ASR Errors
Yik-Cheung Tam, Jiacheng Xu, Jiakai Zou, Zecheng Wang, Tinglong Liao, Shuhan Yuan
Comparative layer-wise analysis of self-supervised speech models
Ankita Pasad, Bowen Shi, Karen Livescu
A Spoken Drug Prescription Dataset in French for Spoken Language Understanding
Ali Can Kocabiyikoglu, François Portet, Prudence Gibert, Hervé Blanchon, Jean-Marc Babouchkine, Gaëtan Gavazzi
End-to-End Spoken Language Understanding: Performance analyses of a voice command task in a low resource setting
Thierry Desot, François Portet, Michel Vacher