Hierarchical Representation
Hierarchical representation learning aims to capture the nested structure of data, mirroring the way humans organize information, to improve model performance and interpretability. Current research focuses on developing novel architectures, such as hierarchical transformers and energy-based models, and algorithms like contrastive learning and variational Bayes, to learn these representations effectively across diverse data types, including images, text, and time series. This work is significant because improved hierarchical representations lead to more robust, efficient, and explainable models with applications ranging from medical image analysis and recommendation systems to robotics and music generation.
Papers
VidMusician: Video-to-Music Generation with Semantic-Rhythmic Alignment via Hierarchical Visual Features
Sifei Li, Binxin Yang, Chunji Yin, Chong Sun, Yuxin Zhang, Weiming Dong, Chen Li
Homogeneous Dynamics Space for Heterogeneous Humans
Xinpeng Liu, Junxuan Liang, Chenshuo Zhang, Zixuan Cai, Cewu Lu, Yong-Lu Li
HARec: Hyperbolic Graph-LLM Alignment for Exploration and Exploitation in Recommender Systems
Qiyao Ma, Menglin Yang, Mingxuan Ju, Tong Zhao, Neil Shah, Rex Ying
CLIPer: Hierarchically Improving Spatial Representation of CLIP for Open-Vocabulary Semantic Segmentation
Lin Sun, Jiale Cao, Jin Xie, Xiaoheng Jiang, Yanwei Pang