Vector Quantized Variational Autoencoder
Vector-quantized variational autoencoders (VQ-VAEs) are generative models that learn discrete representations of data, aiming to achieve efficient compression and high-quality reconstruction or generation. Current research focuses on improving VQ-VAE architectures, such as incorporating decomposed structures, hierarchical refinements, and physics-informed constraints, to enhance performance in diverse applications. These advancements are impacting fields ranging from robotics and speech processing to protein structure prediction and wireless communication, enabling improved data efficiency, generation of realistic data, and more effective anomaly detection.
Papers
CrowdSurfer: Sampling Optimization Augmented with Vector-Quantized Variational AutoEncoder for Dense Crowd Navigation
Naman Kumar, Antareep Singha, Laksh Nanwani, Dhruv Potdar, Tarun R, Fatemeh Rastgar, Simon Idoko, Arun Kumar Singh, K. Madhava Krishna
Exploring VQ-VAE with Prosody Parameters for Speaker Anonymization
Sotheara Leang (CADT, M-PSI), Anderson Augusma (M-PSI, SVH), Eric Castelli (M-PSI), Frédérique Letué (SAM), Sethserey Sam (CADT), Dominique Vaufreydaz (M-PSI)