Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
ChatRex: Taming Multimodal LLM for Joint Perception and Understanding
Qing Jiang, Gen luo, Yuqin Yang, Yuda Xiong, Yihao Chen, Zhaoyang Zeng, Tianhe Ren, Lei Zhang
Can LLMs assist with Ambiguity? A Quantitative Evaluation of various Large Language Models on Word Sense Disambiguation
T.G.D.K. Sumanathilaka, Nicholas Micallef, Julian Hough
Energy-Efficient Split Learning for Fine-Tuning Large Language Models in Edge Networks
Zuguang Li, Shaohua Wu, Liang Li, Songge Zhang
Break the ID-Language Barrier: An Adaption Framework for Sequential Recommendation
Xiaohan Yu, Li Zhang, Xin Zhao, Yue Wang
Evaluating and Improving the Robustness of Security Attack Detectors Generated by LLMs
Samuele Pasini, Jinhan Kim, Tommaso Aiello, Rocio Cabrera Lozoya, Antonino Sabetta, Paolo Tonella
SentiXRL: An advanced large language Model Framework for Multilingual Fine-Grained Emotion Classification in Complex Text Environment
Jie Wang, Yichen Wang, Zhilin Zhang, Jianhao Zeng, Kaidi Wang, Zhiyang Chen
A survey on cutting-edge relation extraction techniques based on language models
Jose A. Diaz-Garcia, Julio Amador Diaz Lopez
Training and Evaluating Language Models with Template-based Data Generation
Yifan Zhang
DRS: Deep Question Reformulation With Structured Output
Zhecheng Li, Yiwei Wang, Bryan Hooi, Yujun Cai, Nanyun Peng, Kai-Wei Chang
Regularized Multi-LLMs Collaboration for Enhanced Score-based Causal Discovery
Xiaoxuan Li, Yao Liu, Ruoyu Wang, Lina Yao
Can LLMs plan paths in the real world?
Wanyi Chen, Meng-Wen Su, Nafisa Mehjabin, Mary L. Cummings
Enhancing Character-Level Understanding in LLMs through Token Internal Structure Learning
Zhu Xu, Zhiqiang Zhao, Zihan Zhang, Yuchi Liu, Quanwei Shen, Fei Liu, Yu Kuang, Jian He, Conglin Liu
Push the Limit of Multi-modal Emotion Recognition by Prompting LLMs with Receptive-Field-Aware Attention Weighting
Liyun Zhang, Dian Ding, Yu Lu, Yi-Chao Chen, Guangtao Xue
MALMM: Multi-Agent Large Language Models for Zero-Shot Robotics Manipulation
Harsh Singh, Rocktim Jyoti Das, Mingfei Han, Preslav Nakov, Ivan Laptev
Pushing the Limits of Large Language Model Quantization via the Linearity Theorem
Vladimir Malinovskii, Andrei Panferov, Ivan Ilin, Han Guo, Peter Richtárik, Dan Alistarh
One Mind, Many Tongues: A Deep Dive into Language-Agnostic Knowledge Neurons in Large Language Models
Pengfei Cao, Yuheng Chen, Zhuoran Jin, Yubo Chen, Kang Liu, Jun Zhao
The Extractive-Abstractive Spectrum: Uncovering Verifiability Trade-offs in LLM Generations
Theodora Worledge, Tatsunori Hashimoto, Carlos Guestrin
Different Bias Under Different Criteria: Assessing Bias in LLMs with a Fact-Based Approach
Changgeon Ko, Jisu Shin, Hoyun Song, Jeongyeon Seo, Jong C. Park
Meaningless is better: hashing bias-inducing words in LLM prompts improves performance in logical reasoning and statistical learning
Milena Chadimová, Eduard Jurášek, Tomáš Kliegr
Using Large Language Models for Expert Prior Elicitation in Predictive Modelling
Alexander Capstick, Rahul G. Krishnan, Payam Barnaghi