Natural Language Processing
Natural Language Processing (NLP) focuses on enabling computers to understand, interpret, and generate human language. Current research heavily emphasizes large language models (LLMs), exploring their capabilities in various tasks like question answering, text classification, and translation, while also addressing challenges such as bias, efficiency, and the need for better evaluation metrics. The field's significance lies in its potential to revolutionize numerous applications, from improving healthcare and education to enhancing information access and facilitating more effective human-computer interaction.
Papers
Unveiling Topological Structures in Text: A Comprehensive Survey of Topological Data Analysis Applications in NLP
Adaku Uchendu, Thai Le
Multidimensional Byte Pair Encoding: Shortened Sequences for Improved Visual Data Generation
Tim Elsner, Paula Usinger, Julius Nehring-Wirxel, Gregor Kobsik, Victor Czech, Yanjiang He, Isaak Lim, Leif Kobbelt
Information Extraction from Clinical Notes: Are We Ready to Switch to Large Language Models?
Yan Hu, Xu Zuo, Yujia Zhou, Xueqing Peng, Jimin Huang, Vipina K. Keloth, Vincent J. Zhang, Ruey-Ling Weng, Qingyu Chen, Xiaoqian Jiang, Kirk E. Roberts, Hua Xu
Initial Nugget Evaluation Results for the TREC 2024 RAG Track with the AutoNuggetizer Framework
Ronak Pradeep, Nandan Thakur, Shivani Upadhyay, Daniel Campos, Nick Craswell, Jimmy Lin
A Practical Guide to Fine-tuning Language Models with Limited Data
Márton Szép, Daniel Rueckert, Rüdiger von Eisenhart-Rothe, Florian Hinterwimmer
DriveThru: a Document Extraction Platform and Benchmark Datasets for Indonesian Local Language Archives
MohammadRifqi Farhansyah, Muhammad Zuhdi Fikri Johari, Afinzaki Amiral, Ayu Purwarianti, Kumara Ari Yuana, Derry Tanti Wijaya
P-MMEval: A Parallel Multilingual Multitask Benchmark for Consistent Evaluation of LLMs
Yidan Zhang, Boyi Deng, Yu Wan, Baosong Yang, Haoran Wei, Fei Huang, Bowen Yu, Junyang Lin, Fei Huang, Jingren Zhou
Virtual teaching assistant for undergraduate students using natural language processing & deep learning
Sadman Jashim Sakib, Baktiar Kabir Joy, Zahin Rydha, Md. Nuruzzaman, Annajiat Alim Rasel
CoCoP: Enhancing Text Classification with LLM through Code Completion Prompt
Mohammad Mahdi Mohajeri, Mohammad Javad Dousti, Majid Nili Ahmadabadi
CamemBERT 2.0: A Smarter French Language Model Aged to Perfection
Wissam Antoun, Francis Kulumba, Rian Touchent, Éric de la Clergerie, Benoît Sagot, Djamé Seddah
1-800-SHARED-TASKS @ NLU of Devanagari Script Languages: Detection of Language, Hate Speech, and Targets using LLMs
Jebish Purbey, Siddartha Pullakhandam, Kanwal Mehreen, Muhammad Arham, Drishti Sharma, Ashay Srivastava, Ram Mohan Rao Kadiyala
AssistRAG: Boosting the Potential of Large Language Models with an Intelligent Information Assistant
Yujia Zhou, Zheng Liu, Zhicheng Dou
A Comprehensive Survey and Guide to Multimodal Large Language Models in Vision-Language Tasks
Chia Xin Liang, Pu Tian, Caitlyn Heqi Yin, Yao Yua, Wei An-Hou, Li Ming, Tianyang Wang, Ziqian Bi, Ming Liu
Building an Efficient Multilingual Non-Profit IR System for the Islamic Domain Leveraging Multiprocessing Design in Rust
Vera Pavlova, Mohammed Makhlouf
The Empirical Impact of Data Sanitization on Language Models
Anwesan Pal, Radhika Bhargava, Kyle Hinsz, Jacques Esterhuizen, Sudipta Bhattacharya
How Good is Your Wikipedia?
Kushal Tatariya, Artur Kulmizev, Wessel Poelman, Esther Ploeger, Marcel Bollmann, Johannes Bjerva, Jiaming Luo, Heather Lent, Miryam de Lhoneux
KyrgyzNLP: Challenges, Progress, and Future
Anton Alekseev, Timur Turatali