Multi Task Learning
Multi-task learning (MTL) aims to improve model efficiency and generalization by training a single model to perform multiple related tasks simultaneously. Current research focuses on addressing challenges like task interference and optimization difficulties, exploring architectures such as Mixture-of-Experts (MoE), low-rank adaptors, and hierarchical models to enhance performance and efficiency across diverse tasks. MTL's significance lies in its potential to improve resource utilization and create more robust and adaptable AI systems, with applications spanning various fields including natural language processing, computer vision, and scientific modeling.
Papers
Entity Tracking via Effective Use of Multi-Task Learning Model and Mention-guided Decoding
Janvijay Singh, Fan Bai, Zhen Wang
Semantic Cross Attention for Few-shot Learning
Bin Xiao, Chien-Liang Liu, Wen-Hoar Hsaio
Task Compass: Scaling Multi-task Pre-training with Task Prefix
Zhuosheng Zhang, Shuohang Wang, Yichong Xu, Yuwei Fang, Wenhao Yu, Yang Liu, Hai Zhao, Chenguang Zhu, Michael Zeng
Optimizing Evaluation Metrics for Multi-Task Learning via the Alternating Direction Method of Multipliers
Ge-Yang Ke, Yan Pan, Jian Yin, Chang-Qin Huang
Multi-Task Dynamical Systems
Alex Bird, Christopher K. I. Williams, Christopher Hawthorne
Improving End-to-End Text Image Translation From the Auxiliary Text Translation Task
Cong Ma, Yaping Zhang, Mei Tu, Xu Han, Linghui Wu, Yang Zhao, Yu Zhou
Data-Efficiency with a Single GPU: An Exploration of Transfer Methods for Small Language Models
Alon Albalak, Akshat Shrivastava, Chinnadhurai Sankar, Adithya Sagar, Mike Ross