Residual Transformer
Residual Transformers combine the strengths of Transformer networks, particularly their ability to capture long-range dependencies, with residual connections to improve training stability and performance. Current research focuses on adapting this architecture for various applications, including medical image analysis (e.g., generating ADC maps, brain tumor segmentation), time series forecasting, and speech processing, often incorporating techniques like low-rank weight sharing to reduce model size and computational cost. These advancements are significant because they enable the deployment of powerful Transformer models in resource-constrained environments and improve accuracy across a range of important tasks.
Papers
November 19, 2024
October 31, 2024
July 2, 2024
May 6, 2024
October 3, 2023
September 28, 2023
July 31, 2023
June 6, 2023
April 28, 2023
March 13, 2023
October 12, 2022
May 19, 2022