Cross Lingual Performance
Cross-lingual performance in large language models (LLMs) focuses on improving the ability of these models to understand and generate text across multiple languages, particularly addressing the challenges posed by low-resource languages. Current research emphasizes techniques like continual pre-training on massive multilingual datasets, efficient fine-tuning methods (e.g., simplified RAFT), and prompt engineering strategies to enhance zero-shot cross-lingual transfer. These advancements are crucial for broadening the accessibility and applicability of NLP technologies globally, fostering linguistic inclusivity and enabling more effective cross-cultural communication and information processing.
Papers
1+1>2: Can Large Language Models Serve as Cross-Lingual Knowledge Aggregators?
Yue Huang, Chenrui Fan, Yuan Li, Siyuan Wu, Tianyi Zhou, Xiangliang Zhang, Lichao Sun
MultiTalk: Enhancing 3D Talking Head Generation Across Languages with Multilingual Video Dataset
Kim Sung-Bin, Lee Chae-Yeon, Gihun Son, Oh Hyun-Bin, Janghoon Ju, Suekyeong Nam, Tae-Hyun Oh