Low Resource
Low-resource settings in natural language processing and related fields present significant challenges due to limited data and computational resources. Current research focuses on adapting existing large language models (LLMs) and other deep learning architectures, such as U-Net and transformer models, through techniques like parameter-efficient fine-tuning, data augmentation (including back-translation and synthetic data generation), and cross-lingual transfer learning to improve performance in tasks such as machine translation, speech recognition, and sentiment analysis for under-resourced languages. These advancements are crucial for bridging the digital divide and enabling access to AI-powered tools and services for a wider range of languages and communities.
Papers
CANTONMT: Investigating Back-Translation and Model-Switch Mechanisms for Cantonese-English Neural Machine Translation
Kung Yin Hong, Lifeng Han, Riza Batista-Navarro, Goran Nenadic
Rethinking Histology Slide Digitization Workflows for Low-Resource Settings
Talat Zehra, Joseph Marino, Wendy Wang, Grigoriy Frantsuzov, Saad Nadeem
ANGOFA: Leveraging OFA Embedding Initialization and Synthetic Data for Angolan Language Model
Osvaldo Luamba Quinjica, David Ifeoluwa Adelani
Enhancing Low-Resource LLMs Classification with PEFT and Synthetic Data
Parth Patwa, Simone Filice, Zhiyu Chen, Giuseppe Castellucci, Oleg Rokhlenko, Shervin Malmasi