Individual Representation
Individual representation in machine learning focuses on creating effective data representations that capture essential information for downstream tasks, improving model performance and robustness. Current research emphasizes developing novel representation learning methods using architectures like variational autoencoders, transformers, and graph neural networks, often incorporating contrastive learning or self-supervised techniques to handle diverse data types and address challenges like non-IID data and co-occurrence issues. These advancements are crucial for various applications, including image generation, object detection, user modeling, and scientific discovery, by enabling more accurate and efficient models across diverse domains. The development of robust and generalizable representations is a key challenge driving ongoing research.
Papers
GDiffRetro: Retrosynthesis Prediction with Dual Graph Enhanced Molecular Representation and Diffusion Generation
Shengyin Sun, Wenhao Yu, Yuxiang Ren, Weitao Du, Liwei Liu, Xuecang Zhang, Ying Hu, Chen Ma
Robust Hyperspectral Image Panshapring via Sparse Spatial-Spectral Representation
Chia-Ming Lee, Yu-Fan Lin, Li-Wei Kang, Chih-Chung Hsu
Spiking Neural Network Accelerator Architecture for Differential-Time Representation using Learned Encoding
Daniel Windhager, Lothar Ratschbacher, Bernhard A. Moser, Michael Lunglmayr