Transformer Based
Transformer-based models are revolutionizing various fields by leveraging self-attention mechanisms to capture long-range dependencies in sequential data, achieving state-of-the-art results in tasks ranging from natural language processing and image recognition to time series forecasting and robotic control. Current research focuses on improving efficiency (e.g., through quantization and optimized architectures), enhancing generalization capabilities, and addressing challenges like handling long sequences and endogeneity. These advancements are significantly impacting diverse scientific communities and practical applications, leading to more accurate, efficient, and robust models across numerous domains.
Papers
AttEntropy: Segmenting Unknown Objects in Complex Scenes using the Spatial Attention Entropy of Semantic Segmentation Transformers
Krzysztof Lis, Matthias Rottmann, Sina Honari, Pascal Fua, Mathieu Salzmann
On Transforming Reinforcement Learning by Transformer: The Development Trajectory
Shengchao Hu, Li Shen, Ya Zhang, Yixin Chen, Dacheng Tao
Beyond SOT: Tracking Multiple Generic Objects at Once
Christoph Mayer, Martin Danelljan, Ming-Hsuan Yang, Vittorio Ferrari, Luc Van Gool, Alina Kuznetsova
Ultra-High-Definition Low-Light Image Enhancement: A Benchmark and Transformer-Based Method
Tao Wang, Kaihao Zhang, Tianrun Shen, Wenhan Luo, Bjorn Stenger, Tong Lu
Assessing the Capacity of Transformer to Abstract Syntactic Representations: A Contrastive Analysis Based on Long-distance Agreement
Bingzhi Li, Guillaume Wisniewski, Benoît Crabbé
NRTR: Neuron Reconstruction with Transformer from 3D Optical Microscopy Images
Yijun Wang, Rui Lang, Rui Li, Junsong Zhang
IncepFormer: Efficient Inception Transformer with Pyramid Pooling for Semantic Segmentation
Lihua Fu, Haoyue Tian, Xiangping Bryce Zhai, Pan Gao, Xiaojiang Peng
Open World DETR: Transformer based Open World Object Detection
Na Dong, Yongqiang Zhang, Mingli Ding, Gim Hee Lee
A K-variate Time Series Is Worth K Words: Evolution of the Vanilla Transformer Architecture for Long-term Multivariate Time Series Forecasting
Zanwei Zhou, Ruizhe Zhong, Chen Yang, Yan Wang, Xiaokang Yang, Wei Shen