Multi Task Model
Multi-task models aim to train a single model capable of performing multiple tasks simultaneously, improving efficiency and generalization compared to training separate models for each task. Current research focuses on developing effective architectures and algorithms, including transformer-based models, mixture-of-experts, and various model merging techniques like task arithmetic and weight averaging, to address challenges such as catastrophic forgetting and representation bias. This field is significant because it offers improved resource utilization and enhanced performance across diverse applications, ranging from medical image analysis and natural language processing to robotics and recommender systems.
Papers
Multi-Task Learning for Fatigue Detection and Face Recognition of Drivers via Tree-Style Space-Channel Attention Fusion Network
Shulei Qu, Zhenguo Gao, Xiaowei Chen, Na Li, Yakai Wang, Xiaoxiao Wu
Localizing Task Information for Improved Model Merging and Compression
Ke Wang, Nikolaos Dimitriadis, Guillermo Ortiz-Jimenez, François Fleuret, Pascal Frossard