Lightweight Transformer
Lightweight Transformers aim to reduce the computational cost and memory footprint of standard Transformer architectures while maintaining performance comparable to their larger counterparts. Current research focuses on developing efficient model variations, such as those incorporating sparse attention mechanisms, hybrid CNN-Transformer designs, and unrolled optimization algorithms, often applied to tasks like image processing, natural language processing, and time-series analysis. This pursuit of efficiency is crucial for deploying Transformer models on resource-constrained devices and expanding their applicability across diverse fields, including medical imaging, mobile computing, and embedded systems.
Papers
Traffic Signal Control Using Lightweight Transformers: An Offline-to-Online RL Approach
Xingshuai Huang, Di Wu, Benoit Boulet
Interfacing Foundation Models' Embeddings
Xueyan Zou, Linjie Li, Jianfeng Wang, Jianwei Yang, Mingyu Ding, Zhengyuan Yang, Feng Li, Hao Zhang, Shilong Liu, Arul Aravinthan, Yong Jae Lee, Lijuan Wang