Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
FlashSloth: Lightning Multimodal Large Language Models via Embedded Visual Compression
Bo Tong, Bokai Lai, Yiyi Zhou, Gen Luo, Yunhang Shen, Ke Li, Xiaoshuai Sun, Rongrong Ji
Densing Law of LLMs
Chaojun Xiao, Jie Cai, Weilin Zhao, Guoyang Zeng, Xu Han, Zhiyuan Liu, Maosong Sun
ALMA: Alignment with Minimal Annotation
Michihiro Yasunaga, Leonid Shamis, Chunting Zhou, Andrew Cohen, Jason Weston, Luke Zettlemoyer, Marjan Ghazvininejad
Evolutionary Pre-Prompt Optimization for Mathematical Reasoning
Mathurin Videau, Alessandro Leite, Marc Schoenauer, Olivier Teytaud
AL-QASIDA: Analyzing LLM Quality and Accuracy Systematically in Dialectal Arabic
Nathaniel R. Robinson, Shahd Abdelmoneim, Kelly Marchisio, Sebastian Ruder
Leveraging Large Language Models to Generate Course-specific Semantically Annotated Learning Objects
Dominic Lohr, Marc Berges, Abhishek Chugh, Michael Kohlhase, Dennis Müller
Monet: Mixture of Monosemantic Experts for Transformers
Jungwoo Park, Young Jin Ahn, Kee-Eung Kim, Jaewoo Kang
Pre-train, Align, and Disentangle: Empowering Sequential Recommendation with Large Language Models
Yuhao Wang, Junwei Pan, Xiangyu Zhao, Pengyue Jia, Wanyu Wang, Yuan Wang, Yue Liu, Dapeng Liu, Jie Jiang
Guidance is All You Need: Temperature-Guided Reasoning in Large Language Models
Eyad Gomaa, Gomaa Salah
Practical Considerations for Agentic LLM Systems
Chris Sypherd, Vaishak Belle
LossAgent: Towards Any Optimization Objectives for Image Processing with LLM Agents
Bingchen Li, Xin Li, Yiting Lu, Zhibo Chen
MTMT: Consolidating Multiple Thinking Modes to Form a Thought Tree for Strengthening LLM
Changcheng Li, Xiangyu Wang, Qiuju Chen, Xiren Zhou, Huanhuan Chen
Chain-of-Thought in Large Language Models: Decoding, Projection, and Activation
Hao Yang, Qianghua Zhao, Lei Li
A Survey on Large Language Model-Based Social Agents in Game-Theoretic Scenarios
Xiachong Feng, Longxu Dou, Ella Li, Qinghao Wang, Haochuan Wang, Yu Guo, Chang Ma, Lingpeng Kong
Uniform Discretized Integrated Gradients: An effective attribution based method for explaining large language models
Swarnava Sinha Roy, Ayan Kundu
Beyond the Binary: Capturing Diverse Preferences With Reward Regularization
Vishakh Padmakumar, Chuanyang Jin, Hannah Rose Kirk, He He
Prompting Large Language Models for Clinical Temporal Relation Extraction
Jianping He, Laila Rasmy, Haifang Li, Jianfu Li, Zenan Sun, Evan Yu, Degui Zhi, Cui Tao
A Review on Scientific Knowledge Extraction using Large Language Models in Biomedical Sciences
Gabriel Lino Garcia, João Renato Ribeiro Manesco, Pedro Henrique Paiola, Lucas Miranda, Maria Paola de Salvo, João Paulo Papa
Chatting with Logs: An exploratory study on Finetuning LLMs for LogQL
Vishwanath Seshagiri, Siddharth Balyan, Vaastav Anand, Kaustubh Dhole, Ishan Sharma, Avani Wildani, José Cambronero, Andreas Züfle
Alignment at Pre-training! Towards Native Alignment for Arabic LLMs
Juhao Liang, Zhenyang Cai, Jianqing Zhu, Huang Huang, Kewei Zong, Bang An, Mosen Alharthi, Juncai He, Lian Zhang, Haizhou Li, Benyou Wang, Jinchao Xu