Low Resource
Low-resource settings in natural language processing and related fields present significant challenges due to limited data and computational resources. Current research focuses on adapting existing large language models (LLMs) and other deep learning architectures, such as U-Net and transformer models, through techniques like parameter-efficient fine-tuning, data augmentation (including back-translation and synthetic data generation), and cross-lingual transfer learning to improve performance in tasks such as machine translation, speech recognition, and sentiment analysis for under-resourced languages. These advancements are crucial for bridging the digital divide and enabling access to AI-powered tools and services for a wider range of languages and communities.
Papers
On Importance of Layer Pruning for Smaller BERT Models and Low Resource Languages
Mayur Shirke, Amey Shembade, Madhushri Wagh, Pavan Thorat, Raviraj Joshi
CODEOFCONDUCT at Multilingual Counterspeech Generation: A Context-Aware Model for Robust Counterspeech Generation in Low-Resource Languages
Michael Bennie, Bushi Xiao, Chryseis Xinyi Liu, Demi Zhang, Jian Meng, Alayo Tripp
Overview of the First Workshop on Language Models for Low-Resource Languages (LoResLM 2025)
Hansi Hettiarachchi, Tharindu Ranasinghe, Paul Rayson, Ruslan Mitkov, Mohamed Gaber, Damith Premasiri, Fiona Anting Tan, Lasitha Uyangodage
Deciphering the Underserved: Benchmarking LLM OCR for Low-Resource Scripts
Muhammad Abdullah Sohail, Salaar Masood, Hamza Iqbal
Multi-OphthaLingua: A Multilingual Benchmark for Assessing and Debiasing LLM Ophthalmological QA in LMICs
David Restrepo, Chenwei Wu, Zhengxu Tang, Zitao Shuai, Thao Nguyen Minh Phan, Jun-En Ding, Cong-Tinh Dao, Jack Gallifant, Robyn Gayle Dychiao, Jose Carlo Artiaga, André Hiroshi Bando, Carolina Pelegrini Barbosa Gracitelli, Vincenz Ferrer, Leo Anthony Celi, Danielle Bitterman, Michael G Morley, Luis Filipe Nakayama
Pipeline Analysis for Developing Instruct LLMs in Low-Resource Languages: A Case Study on Basque
Ander Corral, Ixak Sarasua, Xabier Saralegi
Domain-adaptative Continual Learning for Low-resource Tasks: Evaluation on Nepali
Sharad Duwal, Suraj Prasai, Suresh Manandhar
Cross-Dialect Information Retrieval: Information Access in Low-Resource and High-Variance Languages
Robert Litschko, Oliver Kraus, Verena Blaschke, Barbara Plank
LinguaLIFT: An Effective Two-stage Instruction Tuning Framework for Low-Resource Language Tasks
Hongbin Zhang, Kehai Chen, Xuefeng Bai, Yang Xiang, Min Zhang
Bridging the Gap: Enhancing LLM Performance for Low-Resource African Languages with New Benchmarks, Fine-Tuning, and Cultural Adjustments
Tuka Alhanai, Adam Kasumovic, Mohammad Ghassemi, Aven Zitzelberger, Jessica Lundin, Guillaume Chabot-Couture
PICLe: Pseudo-Annotations for In-Context Learning in Low-Resource Named Entity Detection
Sepideh Mamooler, Syrielle Montariol, Alexander Mathis, Antoine Bosselut