Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Identifying and Manipulating Personality Traits in LLMs Through Activation Engineering
Rumi A. Allbert, James K. Wiles, Vlad Grankovsky
How to Choose a Threshold for an Evaluation Metric for Large Language Models
Bhaskarjit Sarmah, Mingshu Li, Jingrao Lyu, Sebastian Frank, Nathalia Castellanos, Stefano Pasquali, Dhagash Mehta
Asking Again and Again: Exploring LLM Robustness to Repeated Questions
Sagi Shaier
Low-Rank Correction for Quantized LLMs
Meyer Scetbon, James Hensman
FlexLLM: Exploring LLM Customization for Moving Target Defense on Black-Box LLMs Against Jailbreak Attacks
Bocheng Chen, Hanqing Guo, Qiben Yan
DRUM: Learning Demonstration Retriever for Large MUlti-modal Models
Ellen Yi-Ge, Jiechao Gao, Wei Han, Wei Zhu
Adapting to Non-Stationary Environments: Multi-Armed Bandit Enhanced Retrieval-Augmented Generation on Knowledge Graphs
Xiaqiang Tang, Jian Li, Nan Du, Sihong Xie
LLM-as-an-Interviewer: Beyond Static Testing Through Dynamic LLM Evaluation
Eunsu Kim, Juyoung Suk, Seungone Kim, Niklas Muennighoff, Dongkwan Kim, Alice Oh
Hyperband-based Bayesian Optimization for Black-box Prompt Selection
Lennart Schneider, Martin Wistuba, Aaron Klein, Jacek Golebiowski, Giovanni Zappella, Felice Antonio Merra
CoPrUS: Consistency Preserving Utterance Synthesis towards more realistic benchmark dialogues
Sebastian Steindl, Ulrich Schäfer, Bernd Ludwig
Look Before You Leap: Enhancing Attention and Vigilance Regarding Harmful Content with GuidelineLLM
Shaoqing Zhang, Zhuosheng Zhang, Kehai Chen, Rongxiang Weng, Muyun Yang, Tiejun Zhao, Min Zhang
Optimizing Alignment with Less: Leveraging Data Augmentation for Personalized Evaluation
Javad Seraj, Mohammad Mahdi Mohajeri, Mohammad Javad Dousti, Majid Nili Ahmadabadi
Generating Knowledge Graphs from Large Language Models: A Comparative Study of GPT-4, LLaMA 2, and BERT
Ahan Bhatt, Nandan Vaghela, Kush Dudhia
Na'vi or Knave: Jailbreaking Language Models via Metaphorical Avatars
Yu Yan, Sheng Sun, Junqi Tong, Min Liu, Qi Li
ConceptSearch: Towards Efficient Program Search Using LLMs for Abstraction and Reasoning Corpus (ARC)
Kartik Singhal, Gautam Shroff
The Rise and Down of Babel Tower: Investigating the Evolution Process of Multilingual Code Large Language Model
Jiawei Chen, Wentao Chen, Jing Su, Jingjing Xu, Hongyu Lin, Mengjie Ren, Yaojie Lu, Xianpei Han, Le Sun
Enhancing Relation Extraction via Supervised Rationale Verification and Feedback
Yongqi Li, Xin Miao, Shen Zhou, Mayi Xu, Yuyang Ren, Tieyun Qian
HARP: Hesitation-Aware Reframing in Transformer Inference Pass
Romain Storaï, Seung-won Hwang
MemHunter: Automated and Verifiable Memorization Detection at Dataset-scale in LLMs
Zhenpeng Wu, Jian Lou, Zibin Zheng, Chuan Chen
KULTURE Bench: A Benchmark for Assessing Language Model in Korean Cultural Context
Xiaonan Wang, Jinyoung Yeo, Joon-Ho Lim, Hansaem Kim