Articulatory Representation
Articulatory representation focuses on modeling the movements of speech articulators (e.g., tongue, lips) and their relationship to speech sounds, aiming to create more natural and interpretable speech processing systems. Current research emphasizes developing robust methods for inferring articulatory parameters from acoustic signals, using techniques like neural networks (including generative adversarial networks and autoencoders), factor graphs, and matrix factorization, often incorporating multimodal data (audio and visual). This work has significant implications for improving speech synthesis, recognition, and the understanding of speech production mechanisms, particularly in challenging scenarios like dysarthric speech or low-resource languages, and for enabling more sophisticated robot-object interaction.
Papers
Articulatory Configurations across Genders and Periods in French Radio and TV archives
Benjamin Elie, David Doukhan, Rémi Uro, Lucas Ondel-Yang, Albert Rilliard, Simon Devauchelle
Simulating Articulatory Trajectories with Phonological Feature Interpolation
Angelo Ortiz Tandazo, Thomas Schatz, Thomas Hueber, Emmanuel Dupoux
Online Estimation of Articulated Objects with Factor Graphs using Vision and Proprioceptive Sensing
Russell Buchanan, Adrian Röfer, João Moura, Abhinav Valada, Sethu Vijayakumar
GAMMA: Generalizable Articulation Modeling and Manipulation for Articulated Objects
Qiaojun Yu, Junbo Wang, Wenhai Liu, Ce Hao, Liu Liu, Lin Shao, Weiming Wang, Cewu Lu
Articulatory Representation Learning Via Joint Factor Analysis and Neural Matrix Factorization
Jiachen Lian, Alan W Black, Yijing Lu, Louis Goldstein, Shinji Watanabe, Gopala K. Anumanchipalli
Learning to Compute the Articulatory Representations of Speech with the MIRRORNET
Yashish M. Siriwardena, Carol Espy-Wilson, Shihab Shamma