Multi Task Transformer
Multi-task transformers leverage the power of transformer architectures to simultaneously learn multiple related tasks from a shared input, improving efficiency and performance compared to training separate models for each task. Current research focuses on enhancing these models through techniques like parameter-efficient fine-tuning, incorporating auxiliary tasks (e.g., contrastive learning), and designing specialized attention mechanisms to manage task interactions. This approach is proving valuable across diverse fields, including medical image analysis, natural language processing, and robotics, by enabling more efficient training and improved generalization across various sub-tasks within a given domain.
Papers
July 1, 2024
June 4, 2024
November 2, 2023
October 7, 2023
August 22, 2023
August 16, 2023
August 14, 2023
May 19, 2023
March 21, 2023
March 20, 2023
November 7, 2022
September 23, 2022
September 12, 2022
August 6, 2022
May 22, 2022
May 17, 2022
April 16, 2022