Sequential Recommendation
Sequential recommendation aims to predict users' next interactions by analyzing their historical behavior sequences, focusing on capturing dynamic preferences and long-term patterns. Current research heavily utilizes transformer-based architectures, large language models (LLMs), and graph neural networks (GNNs), often incorporating techniques like contrastive learning, test-time training, and knowledge distillation to improve accuracy and efficiency, particularly for large-scale datasets. This field is crucial for personalized recommendations in various applications, driving improvements in user experience and business outcomes through more accurate and timely predictions. Addressing challenges like scalability, noise in user data, and the effective integration of diverse data sources (e.g., textual item descriptions, collaborative filtering signals) remains a key focus.
Papers
Enhancing High-order Interaction Awareness in LLM-based Recommender Model
Xinfeng Wang, Jin Cui, Fumiyo Fukumoto, Yoshimi Suzuki
Large Language Model Empowered Embedding Generator for Sequential Recommendation
Qidong Liu, Xian Wu, Wanyu Wang, Yejing Wang, Yuanshao Zhu, Xiangyu Zhao, Feng Tian, Yefeng Zheng