Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
DriVLM: Domain Adaptation of Vision-Language Models in Autonomous Driving
Xuran Zheng, Chang D. Yoo
A Text-Based Knowledge-Embedded Soft Sensing Modeling Approach for General Industrial Process Tasks Based on Large Language Model
Shuo Tong, Han Liu, Runyuan Guo, Xueqiong Tian, Wenqing Wang, Ding Liu, Youmin Zhang
Enhancing Human-Like Responses in Large Language Models
Ethem Yağız Çalık, Talha Rüzgar Akkuş
Demystifying Domain-adaptive Post-training for Financial LLMs
Zixuan Ke, Yifei Ming, Xuan-Phi Nguyen, Caiming Xiong, Shafiq Joty
Step-by-Step Mastery: Enhancing Soft Constraint Following Ability of Large Language Models
Qingyu Ren, Jie Zeng, Qianyu He, Jiaqing Liang, Yanghua Xiao, Weikang Zhou, Zeye Sun, Fei Yu
Rethinking Evaluation of Sparse Autoencoders through the Representation of Polysemous Words
Gouki Minegishi, Hiroki Furuta, Yusuke Iwasawa, Yutaka Matsuo
$\text{Transformer}^2$: Self-adaptive LLMs
Qi Sun, Edoardo Cetin, Yujin Tang
Real-Time Textless Dialogue Generation
Long Mai, Julie Carson-Berndsen
Do Code LLMs Understand Design Patterns?
Zhenyu Pan, Xuefeng Song, Yunkun Wang, Rongyu Cao, Binhua Li, Yongbin Li, Han Liu
Utility-inspired Reward Transformations Improve Reinforcement Learning Training of Language Models
Roberto-Rafael Maura-Rivero, Chirag Nagpal, Roma Patel, Francesco Visin
Efficient and Responsible Adaptation of Large Language Models for Robust and Equitable Top-k Recommendations
Kirandeep Kaur, Manya Chadha, Vinayak Gupta, Chirag Shah
FlairGPT: Repurposing LLMs for Interior Designs
Gabrielle Littlefair, Niladri Shekhar Dutt, Niloy J. Mitra
CGP-Tuning: Structure-Aware Soft Prompt Tuning for Code Vulnerability Detection
Ruijun Feng, Hammond Pearce, Pietro Liguori, Yulei Sui
When LLMs Struggle: Reference-less Translation Evaluation for Low-resource Languages
Archchana Sindhujan, Diptesh Kanojia, Constantin Orasan, Shenbin Qian
Federated Fine-Tuning of LLMs: Framework Comparison and Research Directions
Na Yan, Yang Su, Yansha Deng, Robert Schober
End-to-End Bangla AI for Solving Math Olympiad Problem Benchmark: Leveraging Large Language Model Using Integrated Approach
H.M. Shadman Tabib, Jaber Ahmed Deedar
Who Does the Giant Number Pile Like Best: Analyzing Fairness in Hiring Contexts
Preethi Seshadri, Seraphina Goldfarb-Tarrant
RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Jun Liu, Zhenglun Kong, Peiyan Dong, Xuan Shen, Pu Zhao, Hao Tang, Geng Yuan, Wei Niu, Wenbin Zhang, Xue Lin, Dong Huang, Yanzhi Wang
Scaling Large Language Model Training on Frontier with Low-Bandwidth Partitioning
Lang Xu, Quentin Anthony, Jacob Hatef, Aamir Shafi, Hari Subramoni, Dhabaleswar K. (DK) Panda