Latent Space
Latent space refers to a lower-dimensional representation of high-dimensional data, aiming to capture essential features while reducing computational complexity and improving interpretability. Current research focuses on developing efficient algorithms and model architectures, such as variational autoencoders (VAEs), generative adversarial networks (GANs), and diffusion models, to learn and manipulate these latent spaces for tasks ranging from anomaly detection and image generation to controlling generative models and improving the efficiency of autonomous systems. This work has significant implications across diverse fields, enabling advancements in areas like drug discovery, autonomous driving, and cybersecurity through improved data analysis, model efficiency, and enhanced control over generative processes.
Papers
Estimation of Models with Limited Data by Leveraging Shared Structure
Maryann Rui, Thibaut Horel, Munther Dahleh
Delving into CLIP latent space for Video Anomaly Recognition
Luca Zanella, Benedetta Liberatori, Willi Menapace, Fabio Poiesi, Yiming Wang, Elisa Ricci
SALSA: Semantically-Aware Latent Space Autoencoder
Kathryn E. Kirchoff, Travis Maxfield, Alexander Tropsha, Shawn M. Gomez
Learning Nonparametric High-Dimensional Generative Models: The Empirical-Beta-Copula Autoencoder
Maximilian Coblenz, Oliver Grothe, Fabian Kächele
Looking through the past: better knowledge retention for generative replay in continual learning
Valeriya Khan, Sebastian Cygert, Kamil Deja, Tomasz Trzciński, Bartłomiej Twardowski