Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
BenchmarkCards: Large Language Model and Risk Reporting
Anna Sokol, Nuno Moniz, Elizabeth Daly, Michael Hind, Nitesh Chawla
Self-Pluralising Culture Alignment for Large Language Models
Shaoyang Xu, Yongqi Leng, Linhao Yu, Deyi Xiong
Large Language Models as a Tool for Mining Object Knowledge
Hannah YoungEun An, Lenhart K. Schubert
Facilitating Multi-turn Function Calling for LLMs via Compositional Instruction Tuning
Mingyang Chen, Haoze Sun, Tianpeng Li, Fan Yang, Hao Liang, Keer Lu, Bin Cui, Wentao Zhang, Zenan Zhou, Weipeng Chen
Mechanistic Unlearning: Robust Knowledge Unlearning and Editing via Mechanistic Localization
Phillip Guo, Aaquib Syed, Abhay Sheshadri, Aidan Ewart, Gintare Karolina Dziugaite
Enhancing Mathematical Reasoning in LLMs by Stepwise Correction
Zhenyu Wu, Qingkai Zeng, Zhihan Zhang, Zhaoxuan Tan, Chao Shen, Meng Jiang
Interpreting token compositionality in LLMs: A robustness analysis
Nura Aljaafari, Danilo S. Carvalho, André Freitas
Meta-Chunking: Learning Efficient Text Segmentation via Logical Perception
Jihao Zhao, Zhiyuan Ji, Pengnian Qi, Simin Niu, Bo Tang, Feiyu Xiong, Zhiyu Li
In-Context Learning Enables Robot Action Prediction in LLMs
Yida Yin, Zekai Wang, Yuvan Sharma, Dantong Niu, Trevor Darrell, Roei Herzig
Evaluating Morphological Compositional Generalization in Large Language Models
Mete Ismayilzada, Defne Circi, Jonne Sälevä, Hale Sirin, Abdullatif Köksal, Bhuwan Dhingra, Antoine Bosselut, Lonneke van der Plas, Duygu Ataman
Weak-to-Strong Generalization beyond Accuracy: a Pilot Study in Safety, Toxicity, and Legal Reasoning
Ruimeng Ye, Yang Xiao, Bo Hui
Not All Votes Count! Programs as Verifiers Improve Self-Consistency of Language Models for Math Reasoning
Vernon Y.H. Toh, Deepanway Ghosal, Soujanya Poria
CCSBench: Evaluating Compositional Controllability in LLMs for Scientific Document Summarization
Yixi Ding, Jiaying Wu, Tongyao Zhu, Yanxia Qin, Qian Liu, Min-Yen Kan
On the Utility of Domain Modeling Assistance with Large Language Models
Meriem Ben Chaaben, Lola Burgueño, Istvan David, Houari Sahraoui
FTII-Bench: A Comprehensive Multimodal Benchmark for Flow Text with Image Insertion
Jiacheng Ruan, Yebin Yang, Zehao Lin, Feiyu Xiong, Zeyun Tang, Zhiyu Li
Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse RL
Jared Joselowitz, Arjun Jagota, Satyapriya Krishna, Sonali Parbhoo
Mastering the Craft of Data Synthesis for CodeLLMs
Meng Chen, Philip Arthur, Qianyu Feng, Cong Duy Vu Hoang, Yu-Heng Hong, Mahdi Kazemi Moghaddam, Omid Nezami, Thien Nguyen, Gioacchino Tangari, Duy Vu, Thanh Vu, Mark Johnson, Krishnaram Kenthapadi, Don Dharmasiri, Long Duong, Yuan-Fang Li
KcMF: A Knowledge-compliant Framework for Schema and Entity Matching with Fine-tuning-free LLMs
Yongqin Xu, Huan Li, Ke Chen, Lidan Shou
The Best of Both Worlds: Bridging Quality and Diversity in Data Selection with Bipartite Graph
Minghao Wu, Thuy-Trang Vu, Lizhen Qu, Gholamreza Haffari
Conformity in Large Language Models
Xiaochen Zhu, Caiqi Zhang, Tom Stafford, Nigel Collier, Andreas Vlachos