Dual Learning
Dual learning is a machine learning paradigm that improves model performance by simultaneously training two related but inverse tasks, leveraging the strengths of each to enhance the other. Current research focuses on applying this approach to diverse areas, including reinforcement learning, representation learning for noisy data, and multimodal misinformation detection, often employing contrastive learning, optimal transport, and various adaptation techniques within model architectures like LLMs and VLMs. This framework shows promise for improving efficiency and accuracy in various applications, from enhancing large language models to improving the robustness of weakly supervised learning tasks and even generating more realistic music and dance compositions.
Papers
DUAL-REFLECT: Enhancing Large Language Models for Reflective Translation through Dual Learning Feedback Mechanisms
Andong Chen, Lianzhang Lou, Kehai Chen, Xuefeng Bai, Yang Xiang, Muyun Yang, Tiejun Zhao, Min Zhang
TernaryLLM: Ternarized Large Language Model
Tianqi Chen, Zhe Li, Weixiang Xu, Zeyu Zhu, Dong Li, Lu Tian, Emad Barsoum, Peisong Wang, Jian Cheng