Parallel Text
Parallel text, consisting of texts in multiple languages that are mutual translations, is crucial for machine translation and cross-lingual natural language processing. Current research focuses on improving the efficiency of parallel text acquisition through smart crawling techniques and data augmentation methods, as well as optimizing parallel processing of large language models using techniques like tensor parallelism and kernel fusion to accelerate training and inference. These advancements are vital for bridging language barriers and improving the performance of various NLP applications, particularly for low-resource languages where parallel data is scarce.
Papers
August 19, 2022
July 3, 2022
February 2, 2022
January 18, 2022
December 16, 2021
November 22, 2021