Cross Lingual Transfer
Cross-lingual transfer aims to leverage knowledge learned from high-resource languages to improve performance on low-resource languages in natural language processing tasks. Current research focuses on adapting large language models (LLMs) for cross-lingual transfer, employing techniques like model merging, data augmentation (including synthetic data generation and transliteration), and innovative training strategies such as in-context learning and continual pre-training. This research is crucial for expanding the reach of NLP to a wider range of languages, enabling applications like multilingual question answering, sentiment analysis, and code generation to benefit diverse communities globally.
Papers
Exploring the Relationship between Alignment and Cross-lingual Transfer in Multilingual Transformers
Félix Gaschi, Patricio Cerda, Parisa Rastin, Yannick Toussaint
Cross-Lingual Transfer with Target Language-Ready Task Adapters
Marinela Parović, Alan Ansell, Ivan Vulić, Anna Korhonen
Cross-Lingual Transfer Learning for Phrase Break Prediction with Multilingual Language Model
Hoyeon Lee, Hyun-Wook Yoon, Jong-Hwan Kim, Jae-Min Kim
Improved Cross-Lingual Transfer Learning For Automatic Speech Translation
Sameer Khurana, Nauman Dawalatabad, Antoine Laurent, Luis Vicente, Pablo Gimeno, Victoria Mingote, James Glass
Improving Polish to English Neural Machine Translation with Transfer Learning: Effects of Data Volume and Language Similarity
Juuso Eronen, Michal Ptaszynski, Karol Nowakowski, Zheng Lin Chia, Fumito Masui
The Effects of Input Type and Pronunciation Dictionary Usage in Transfer Learning for Low-Resource Text-to-Speech
Phat Do, Matt Coler, Jelske Dijkstra, Esther Klabbers