Transformer Based
Transformer-based models are revolutionizing various fields by leveraging self-attention mechanisms to capture long-range dependencies in sequential data, achieving state-of-the-art results in tasks ranging from natural language processing and image recognition to time series forecasting and robotic control. Current research focuses on improving efficiency (e.g., through quantization and optimized architectures), enhancing generalization capabilities, and addressing challenges like handling long sequences and endogeneity. These advancements are significantly impacting diverse scientific communities and practical applications, leading to more accurate, efficient, and robust models across numerous domains.
Papers
Overcoming a Theoretical Limitation of Self-Attention
David Chiang, Peter Cholak
A Transformer-based Network for Deformable Medical Image Registration
Yibo Wang, Wen Qian, Xuming Zhang
When Transformer Meets Robotic Grasping: Exploits Context for Efficient Grasp Detection
Shaochen Wang, Zhangli Zhou, Zhen Kan
Transformer for Graphs: An Overview from Architecture Perspective
Erxue Min, Runfa Chen, Yatao Bian, Tingyang Xu, Kangfei Zhao, Wenbing Huang, Peilin Zhao, Junzhou Huang, Sophia Ananiadou, Yu Rong
TraSeTR: Track-to-Segment Transformer with Contrastive Query for Instance-level Instrument Segmentation in Robotic Surgery
Zixu Zhao, Yueming Jin, Pheng-Ann Heng