Adaptive Transformer
Adaptive transformers are a class of neural network models designed to dynamically adjust their behavior based on input data characteristics, improving performance and efficiency across diverse tasks. Current research focuses on developing adaptive attention mechanisms, often incorporating probabilistic methods or Kalman filtering, to selectively weigh input information and handle varying data complexities, such as noise, missing data, or domain shifts. These advancements are impacting various fields, including real-time perception for autonomous systems, survival analysis, and few-shot learning, by enabling more robust and accurate models for challenging applications. The resulting models often demonstrate superior performance compared to traditional transformer architectures, particularly in scenarios with high variability or limited data.
Papers
Transtreaming: Adaptive Delay-aware Transformer for Real-time Streaming Perception
Xiang Zhang, Yufei Cui, Chenchen Fu, Weiwei Wu, Zihao Wang, Yuyang Sun, Xue Liu
Adaptive Transformer Modelling of Density Function for Nonparametric Survival Analysis
Xin Zhang, Deval Mehta, Yanan Hu, Chao Zhu, David Darby, Zhen Yu, Daniel Merlo, Melissa Gresle, Anneke Van Der Walt, Helmut Butzkueven, Zongyuan Ge