Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
AutoFPDesigner: Automated Flight Procedure Design Based on Multi-Agent Large Language Model
Longtao Zhu, Hongyu Yang, Ge Song, Xin Ma, Yanxin Zhang, Yulong Ji
LangGFM: A Large Language Model Alone Can be a Powerful Graph Foundation Model
Tianqianjin Lin, Pengwei Yan, Kaisong Song, Zhuoren Jiang, Yangyang Kang, Jun Lin, Weikang Yuan, Junjie Cao, Changlong Sun, Xiaozhong Liu
SPRIG: Improving Large Language Model Performance by System Prompt Optimization
Lechen Zhang, Tolga Ergen, Lajanugen Logeswaran, Moontae Lee, David Jurgens
SudoLM: Learning Access Control of Parametric Knowledge with Authorization Alignment
Qin Liu, Fei Wang, Chaowei Xiao, Muhao Chen
Enhancing Large Language Models' Situated Faithfulness to External Contexts
Yukun Huang, Sanxing Chen, Hongyi Cai, Bhuwan Dhingra
Distance between Relevant Information Pieces Causes Bias in Long-Context LLMs
Runchu Tian, Yanghao Li, Yuepeng Fu, Siyang Deng, Qinyu Luo, Cheng Qian, Shuo Wang, Xin Cong, Zhong Zhang, Yesai Wu, Yankai Lin, Huadong Wang, Xiaojiang Liu
Do LLMs estimate uncertainty well in instruction-following?
Juyeon Heo, Miao Xiong, Christina Heinze-Deml, Jaya Narain
Large Language Models Are Overparameterized Text Encoders
Thennal D K, Tim Fischer, Chris Biemann
Understanding the difficulty of low-precision post-training quantization of large language models
Zifei Xu, Sayeh Sharify, Wanzin Yazar, Tristan Webb, Xin Wang
When LLMs Go Online: The Emerging Threat of Web-Enabled LLMs
Hanna Kim, Minkyoo Song, Seung Ho Na, Seungwon Shin, Kimin Lee
Tell me what I need to know: Exploring LLM-based (Personalized) Abstractive Multi-Source Meeting Summarization
Frederic Kirstein, Terry Ruas, Robert Kratel, Bela Gipp
Enabling Scalable Evaluation of Bias Patterns in Medical LLMs
Hamed Fayyaz, Raphael Poulain, Rahmatollah Beheshti
Combining Entropy and Matrix Nuclear Norm for Enhanced Evaluation of Language Models
James Vo
A Systematic Study of Cross-Layer KV Sharing for Efficient LLM Inference
You Wu, Haoyi Wu, Kewei Tu
Analyzing Context Utilization of LLMs in Document-Level Translation
Wafaa Mohammed, Vlad Niculae
CoMAL: Collaborative Multi-Agent Large Language Models for Mixed-Autonomy Traffic
Huaiyuan Yao, Longchao Da, Vishnu Nandam, Justin Turnau, Zhiwei Liu, Linsey Pang, Hua Wei
Good Parenting is all you need -- Multi-agentic LLM Hallucination Mitigation
Edward (Ted) Kwartler, Matthew Berman, Alan Aqrawi
Beyond Binary: Towards Fine-Grained LLM-Generated Text Detection via Role Recognition and Involvement Measurement
Zihao Cheng, Li Zhou, Feng Jiang, Benyou Wang, Haizhou Li
Revisiting SLO and Goodput Metrics in LLM Serving
Zhibin Wang, Shipeng Li, Yuhang Zhou, Xue Li, Rong Gu, Nguyen Cam-Tu, Chen Tian, Sheng Zhong
Synthesizing Post-Training Data for LLMs through Multi-Agent Simulation
Shuo Tang, Xianghe Pang, Zexi Liu, Bohan Tang, Rui Ye, Xiaowen Dong, Yanfeng Wang, Siheng Chen