Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
GroverGPT: A Large Language Model with 8 Billion Parameters for Quantum Searching
Haoran Wang, Pingzhi Li, Min Chen, Jinglei Cheng, Junyu Liu, Tianlong Chen
Distributed Mixture-of-Agents for Edge Inference with Large Language Models
Purbesh Mitra, Priyanka Kaswan, Sennur Ulukus
Facilitating large language model Russian adaptation with Learned Embedding Propagation
Mikhail Tikhomirov, Daniil Chernyshev
Plug-and-Play Training Framework for Preference Optimization
Jingyuan Ma, Rui Li, Zheng Li, Lei Sha, Zhifang Sui
KARPA: A Training-free Method of Adapting Knowledge Graph as References for Large Language Model's Reasoning Path Aggregation
Siyuan Fang, Kaijing Ma, Tianyu Zheng, Xinrun Du, Ningxuan Lu, Ge Zhang, Qingkun Tang
Enhancing AI Safety Through the Fusion of Low Rank Adapters
Satya Swaroop Gudipudi, Sreeram Vipparla, Harpreet Singh, Shashwat Goel, Ponnurangam Kumaraguru
DoTA: Weight-Decomposed Tensor Adaptation for Large Language Models
Xiaolin Hu, Xiang Cheng, Peiyu Liu, Wei Liu, Jian Luan, Bin Wang, Yong Liu
Enhancing Annotated Bibliography Generation with LLM Ensembles
Sergio Bermejo
Are LLMs Really Not Knowledgable? Mining the Submerged Knowledge in LLMs' Memory
Xingjian Tao, Yiwei Wang, Yujun Cai, Zhicheng Yang, Jing Tang
SecBench: A Comprehensive Multi-Dimensional Benchmarking Dataset for LLMs in Cybersecurity
Pengfei Jing, Mengyun Tang, Xiaorong Shi, Xing Zheng, Sen Nie, Shi Wu, Yong Yang, Xiapu Luo
Learning to Rank Pre-trained Vision-Language Models for Downstream Tasks
Yuhe Ding, Bo Jiang, Aihua Zheng, Qin Xu, Jian Liang
Align Attention Heads Before Merging Them: An Effective Way for Converting MHA to GQA
Qingyun Jin, Xiaohui Song, Feng Zhou, Zengchang Qin
Knowledge Editing for Large Language Model with Knowledge Neuronal Ensemble
Yongchang Li, Yujin Zhu, Tao Yan, Shijian Fan, Gang Wu, Liang Xu
ReTaKe: Reducing Temporal and Knowledge Redundancy for Long Video Understanding
Xiao Wang, Qingyi Si, Jianlong Wu, Shiyu Zhu, Li Cao, Liqiang Nie
Enhancing Entertainment Translation for Indian Languages using Adaptive Context, Style and LLMs
Pratik Rakesh Singh, Mohammadi Zaki, Pankaj Wasnik
Comparative Performance of Advanced NLP Models and LLMs in Multilingual Geo-Entity Detection
Kalin Kopanov
Multi-Objective Large Language Model Unlearning
Zibin Pan, Shuwen Zhang, Yuesheng Zheng, Chi Li, Yuheng Cheng, Junhua Zhao
Natural Language Fine-Tuning
Jia Liu, Yue Wang, Zhiqi Lin, Min Chen, Yixue Hao, Long Hu
LLM2: Let Large Language Models Harness System 2 Reasoning
Cheng Yang, Chufan Shi, Siheng Li, Bo Shui, Yujiu Yang, Wai Lam