Multi Task Learning
Multi-task learning (MTL) aims to improve model efficiency and generalization by training a single model to perform multiple related tasks simultaneously. Current research focuses on addressing challenges like task interference and optimization difficulties, exploring architectures such as Mixture-of-Experts (MoE), low-rank adaptors, and hierarchical models to enhance performance and efficiency across diverse tasks. MTL's significance lies in its potential to improve resource utilization and create more robust and adaptable AI systems, with applications spanning various fields including natural language processing, computer vision, and scientific modeling.
Papers
Derivative Free Weight-space Ensembling
Dean Ninalga
TBGC: Task-level Backbone-Oriented Gradient Clip for Multi-Task Foundation Model Learning
Zelun Zhang, Xue Pan
Mitigating Negative Transfer with Task Awareness for Sexism, Hate Speech, and Toxic Language Detection
Angel Felipe Magnossão de Paula, Paolo Rosso, Damiano Spina
STG-MTL: Scalable Task Grouping for Multi-Task Learning Using Data Map
Ammar Sherif, Abubakar Abid, Mustafa Elattar, Mohamed ElHelw