Subword Embeddings
Subword embeddings represent words as sequences of smaller units (subwords), improving handling of rare or unseen words in natural language processing. Current research focuses on optimizing subword segmentation algorithms, exploring the interplay between subword representations and cross-lingual transfer in multilingual models, and developing efficient methods for initializing embeddings in low-resource languages. These advancements enhance the performance of various NLP tasks, including machine translation and part-of-speech tagging, particularly for languages with complex morphology or limited training data, and contribute to more efficient and environmentally friendly model training.
Papers
June 19, 2024
March 29, 2024
February 15, 2024
November 15, 2023
May 23, 2023
February 28, 2023
December 19, 2022
December 2, 2022
November 29, 2022
June 7, 2022
May 23, 2022
January 13, 2022
December 20, 2021
December 13, 2021