Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
SC-Phi2: A Fine-tuned Small Language Model for StarCraft II Macromanagement Tasks
Muhammad Junaid Khan, Gita Sukthankar
Strategic Insights in Human and Large Language Model Tactics at Word Guessing Games
Matīss Rikters, Sanita Reinsone
KVPruner: Structural Pruning for Faster and Memory-Efficient Large Language Models
Bo Lv, Quan Zhou, Xuanang Ding, Yan Wang, Zeming Ma
Large Language Models are Good Multi-lingual Learners : When LLMs Meet Cross-lingual Prompts
Teng Wang, Zhenqi He, Wing-Yin Yu, Xiaojin Fu, Xiongwei Han
Hierarchical Narrative Analysis: Unraveling Perceptions of Generative AI
Riona Matsuoka, Hiroki Matsumoto, Takahiro Yoshida, Tomohiro Watanabe, Ryoma Kondo, Ryohei Hisano
Enhancing Multilingual Speech Generation and Recognition Abilities in LLMs with Constructed Code-switched Data
Jing Xu, Daxin Tan, Jiaqi Wang, Xiao Chen
Attention-Seeker: Dynamic Self-Attention Scoring for Unsupervised Keyphrase Extraction
Erwin D. López Z., Cheng Tang, Atsushi Shimada
Adaptive Large Language Models By Layerwise Attention Shortcuts
Prateek Verma, Mert Pilanci
Challenging Fairness: A Comprehensive Exploration of Bias in LLM-Based Recommendations
Shahnewaz Karim Sakib, Anindya Bijoy Das
Model Tells Itself Where to Attend: Faithfulness Meets Automatic Attention Steering
Qingru Zhang, Xiaodong Yu, Chandan Singh, Xiaodong Liu, Liyuan Liu, Jianfeng Gao, Tuo Zhao, Dan Roth, Hao Cheng
Multidimensional Human Activity Recognition With Large Language Model: A Conceptual Framework
Syed Mhamudul Hasan
Model-in-the-Loop (MILO): Accelerating Multimodal AI Data Annotation with LLMs
Yifan Wang, David Stevens, Pranay Shah, Wenwen Jiang, Miao Liu, Xu Chen, Robert Kuo, Na Li, Boying Gong, Daniel Lee, Jiabo Hu, Ning Zhang, Bob Kamma
Do Large Language Models Need a Content Delivery Network?
Yihua Cheng, Kuntai Du, Jiayi Yao, Junchen Jiang
RetrievalAttention: Accelerating Long-Context LLM Inference via Vector Retrieval
Di Liu, Meng Chen, Baotong Lu, Huiqiang Jiang, Zhenhua Han, Qianxi Zhang, Qi Chen, Chengruidong Zhang, Bailu Ding, Kai Zhang, Chen Chen, Fan Yang, Yuqing Yang, Lili Qiu
LLM as BT-Planner: Leveraging LLMs for Behavior Tree Generation in Robot Task Planning
Jicong Ao, Fan Wu, Yansong Wu, Abdalla Swikir, Sami Haddadin
AI Conversational Interviewing: Transforming Surveys with LLMs as Adaptive Interviewers
Alexander Wuttke, Matthias Aßenmacher, Christopher Klamm, Max M. Lang, Quirin Würschinger, Frauke Kreuter
The 20 questions game to distinguish large language models
Gurvan Richardeau, Erwan Le Merrer, Camilla Penzo, Gilles Tredan
Enhancing RL Safety with Counterfactual LLM Reasoning
Dennis Gross, Helge Spieker
Householder Pseudo-Rotation: A Novel Approach to Activation Editing in LLMs with Direction-Magnitude Perspective
Van-Cuong Pham, Thien Huu Nguyen
On the Diagram of Thought
Yifan Zhang, Yang Yuan, Andrew Chi-Chih Yao