General Transformer
General transformers are a class of neural network architectures designed to process sequential data, achieving remarkable success across diverse applications. Current research focuses on improving their generalization capabilities, particularly for longer sequences and unseen data, exploring novel attention mechanisms and model architectures like decoder-only transformers and complementary transformers to enhance efficiency and performance. This work is significant because it addresses limitations in existing models, leading to more robust and adaptable systems with applications ranging from robot control and medical diagnosis to image processing and brain network analysis.
Papers
October 2, 2024
September 24, 2024
July 22, 2024
March 16, 2024
December 8, 2023
October 6, 2023
August 16, 2023
July 23, 2023
June 28, 2023
May 26, 2023
March 11, 2023
October 8, 2022
July 28, 2022
May 30, 2022