Pre Trained Multilingual Model
Pre-trained multilingual models are large language models trained on massive datasets spanning numerous languages, aiming to improve cross-lingual understanding and task performance. Current research focuses on mitigating issues like data bias, improving efficiency (e.g., faster text generation, data-efficient fine-tuning), and enhancing zero-shot cross-lingual transfer capabilities, often employing transformer-based architectures like BERT and its variants. These models are significantly impacting various NLP tasks, from machine translation and named entity recognition to question answering and speech-to-text, particularly benefiting low-resource languages by enabling knowledge transfer from high-resource ones.
Papers
October 11, 2024
August 27, 2024
June 19, 2024
March 30, 2024
March 29, 2024
January 19, 2024
November 22, 2023
November 10, 2023
November 9, 2023
July 6, 2023
May 31, 2023
April 9, 2023
January 16, 2023
December 7, 2022
December 4, 2022
October 13, 2022
August 16, 2022
July 14, 2022
April 29, 2022
December 18, 2021