Transformer Based
Transformer-based models are revolutionizing various fields by leveraging self-attention mechanisms to capture long-range dependencies in sequential data, achieving state-of-the-art results in tasks ranging from natural language processing and image recognition to time series forecasting and robotic control. Current research focuses on improving efficiency (e.g., through quantization and optimized architectures), enhancing generalization capabilities, and addressing challenges like handling long sequences and endogeneity. These advancements are significantly impacting diverse scientific communities and practical applications, leading to more accurate, efficient, and robust models across numerous domains.
Papers
Improving Across-Dataset Brain Tissue Segmentation Using Transformer
Vishwanatha M. Rao, Zihan Wan, Soroush Arabshahi, David J. Ma, Pin-Yu Lee, Ye Tian, Xuzhe Zhang, Andrew F. Laine, Jia Guo
SegTransVAE: Hybrid CNN -- Transformer with Regularization for medical image segmentation
Quan-Dung Pham, Hai Nguyen-Truong, Nam Nguyen Phuong, Khoa N. A. Nguyen
Learning Bounded Context-Free-Grammar via LSTM and the Transformer:Difference and Explanations
Hui Shi, Sicun Gao, Yuandong Tian, Xinyun Chen, Jishen Zhao
TransZero++: Cross Attribute-Guided Transformer for Zero-Shot Learning
Shiming Chen, Ziming Hong, Wenjin Hou, Guo-Sen Xie, Yibing Song, Jian Zhao, Xinge You, Shuicheng Yan, Ling Shao
Block-Skim: Efficient Question Answering for Transformer
Yue Guan, Zhengyi Li, Jingwen Leng, Zhouhan Lin, Minyi Guo, Yuhao Zhu
Trees in transformers: a theoretical analysis of the Transformer's ability to represent trees
Qi He, João Sedoc, Jordan Rodu