Monolingual Pre Trained
Monolingual pre-trained language models focus on training large language models on massive amounts of text from a single language, aiming to achieve superior performance within that language compared to multilingual counterparts. Current research emphasizes efficient training methods for low-resource languages, including techniques like model adaptation and cross-lingual transfer learning, often leveraging architectures such as BERT and LLMs like Llama. This approach is significant because it addresses the performance limitations of multilingual models in specific languages and offers a more cost-effective and sustainable path to developing high-quality language models for a wider range of languages.
Papers
August 21, 2024
May 25, 2024
May 13, 2024
April 12, 2024
April 5, 2024
March 21, 2024
March 19, 2024
November 6, 2023
October 5, 2023
September 22, 2023
April 16, 2023
November 13, 2022
October 10, 2022
May 15, 2022
April 6, 2022
February 28, 2022