Transformer Based
Transformer-based models are revolutionizing various fields by leveraging self-attention mechanisms to capture long-range dependencies in sequential data, achieving state-of-the-art results in tasks ranging from natural language processing and image recognition to time series forecasting and robotic control. Current research focuses on improving efficiency (e.g., through quantization and optimized architectures), enhancing generalization capabilities, and addressing challenges like handling long sequences and endogeneity. These advancements are significantly impacting diverse scientific communities and practical applications, leading to more accurate, efficient, and robust models across numerous domains.
Papers
ETO:Efficient Transformer-based Local Feature Matching by Organizing Multiple Homography Hypotheses
Junjie Ni, Guofeng Zhang, Guanglin Li, Yijin Li, Xinyang Liu, Zhaoyang Huang, Hujun Bao
LoFLAT: Local Feature Matching using Focused Linear Attention Transformer
Naijian Cao, Renjie He, Yuchao Dai, Mingyi He
Context-Scaling versus Task-Scaling in In-Context Learning
Amirhesam Abedsoltan, Adityanarayanan Radhakrishnan, Jingfeng Wu, Mikhail Belkin
Transformer based super-resolution downscaling for regional reanalysis: Full domain vs tiling approaches
Antonio Pérez, Mario Santa Cruz, Daniel San Martín, José Manuel Gutiérrez
What Does It Mean to Be a Transformer? Insights from a Theoretical Hessian Analysis
Weronika Ormaniec, Felix Dangel, Sidak Pal Singh
Domain-Conditioned Transformer for Fully Test-time Adaptation
Yushun Tang, Shuoshuo Chen, Jiyuan Jia, Yi Zhang, Zhihai He
A Consistency-Aware Spot-Guided Transformer for Versatile and Hierarchical Point Cloud Registration
Renlang Huang, Yufan Tang, Jiming Chen, Liang Li
Towards Better Multi-head Attention via Channel-wise Sample Permutation
Shen Yuan, Hongteng Xu