Machine Translation
Machine translation (MT) aims to automatically translate text between languages, with current research heavily focused on leveraging large language models (LLMs) and exploring various architectures like encoder-decoder and decoder-only models. Key areas of investigation include improving translation quality, particularly for low-resource languages and specialized domains like medicine, mitigating biases (e.g., gender bias), and enhancing evaluation methods beyond simple correlation with human judgments. These advancements have significant implications for cross-cultural communication, information access, and the development of more equitable and effective multilingual technologies.
Papers
RAMP: Retrieval and Attribute-Marking Enhanced Prompting for Attribute-Controlled Translation
Gabriele Sarti, Phu Mon Htut, Xing Niu, Benjamin Hsu, Anna Currey, Georgiana Dinu, Maria Nadejde
TranSFormer: Slow-Fast Transformer for Machine Translation
Bei Li, Yi Jing, Xu Tan, Zhen Xing, Tong Xiao, Jingbo Zhu
Do GPTs Produce Less Literal Translations?
Vikas Raunak, Arul Menezes, Matt Post, Hany Hassan Awadalla
Textless Speech-to-Speech Translation With Limited Parallel Data
Anuj Diwan, Anirudh Srinivasan, David Harwath, Eunsol Choi
MuLER: Detailed and Scalable Reference-based Evaluation
Taelin Karidi, Leshem Choshen, Gal Patel, Omri Abend
Leveraging GPT-4 for Automatic Translation Post-Editing
Vikas Raunak, Amr Sharaf, Yiren Wang, Hany Hassan Awadallah, Arul Menezes
CMOT: Cross-modal Mixup via Optimal Transport for Speech Translation
Yan Zhou, Qingkai Fang, Yang Feng
Benchmarking Machine Translation with Cultural Awareness
Binwei Yao, Ming Jiang, Tara Bobinac, Diyi Yang, Junjie Hu
Ties Matter: Meta-Evaluating Modern Metrics with Pairwise Accuracy and Tie Calibration
Daniel Deutsch, George Foster, Markus Freitag
Revisiting Machine Translation for Cross-lingual Classification
Mikel Artetxe, Vedanuj Goswami, Shruti Bhosale, Angela Fan, Luke Zettlemoyer
Accessing Higher Dimensions for Unsupervised Word Translation
Sida I. Wang
CTQScorer: Combining Multiple Features for In-context Example Selection for Machine Translation
Aswanth Kumar, Ratish Puduppully, Raj Dabre, Anoop Kunchukuttan
TeCS: A Dataset and Benchmark for Tense Consistency of Machine Translation
Yiming Ai, Zhiwei He, Kai Yu, Rui Wang
CombLM: Adapting Black-Box Language Models through Small Fine-Tuned Models
Aitor Ormazabal, Mikel Artetxe, Eneko Agirre
Syntactic Knowledge via Graph Attention with BERT in Machine Translation
Yuqian Dai, Serge Sharoff, Marc de Kamps
Improving Isochronous Machine Translation with Target Factors and Auxiliary Counters
Proyag Pal, Brian Thompson, Yogesh Virkar, Prashant Mathur, Alexandra Chronopoulou, Marcello Federico
Extrapolating Multilingual Understanding Models as Multilingual Generators
Bohong Wu, Fei Yuan, Hai Zhao, Lei Li, Jingjing Xu
Decomposed Prompting for Machine Translation Between Related Languages using Large Language Models
Ratish Puduppully, Anoop Kunchukuttan, Raj Dabre, Ai Ti Aw, Nancy F. Chen
Kanbun-LM: Reading and Translating Classical Chinese in Japanese Methods by Language Models
Hao Wang, Hirofumi Shimizu, Daisuke Kawahara