Machine Translated
Machine translation (MT) research currently focuses on improving the accuracy, fluency, and cultural sensitivity of translations, particularly for low-resource languages. This involves investigating model architectures like transformers and exploring techniques to mitigate issues like "translationese" and hallucinations, often through interpretability methods that analyze model internal representations and attention mechanisms. These advancements are crucial for bridging language barriers in various applications, from multilingual information access to cross-cultural communication and improving the quality of training data for other AI models. Furthermore, research emphasizes developing methods to assess and improve the quality of MT outputs, including the use of both human and machine-generated evaluations.
Papers
Focus on Your Question! Interpreting and Mitigating Toxic CoT Problems in Commonsense Reasoning
Jiachun Li, Pengfei Cao, Chenhao Wang, Zhuoran Jin, Yubo Chen, Daojian Zeng, Kang Liu, Jun Zhao
Cutting Off the Head Ends the Conflict: A Mechanism for Interpreting and Mitigating Knowledge Conflicts in Language Models
Zhuoran Jin, Pengfei Cao, Hongbang Yuan, Yubo Chen, Jiexin Xu, Huaijun Li, Xiaojian Jiang, Kang Liu, Jun Zhao
Towards Interpreting Multi-Objective Feature Associations
Nisha Pillai, Ganga Gireesan, Michael J. Rothrock, Bindu Nanduri, Zhiqian Chen, Mahalingam Ramkumar