Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
LLM4AD: A Platform for Algorithm Design with Large Language Model
Fei Liu, Rui Zhang, Zhuoliang Xie, Rui Sun, Kai Li, Xi Lin, Zhenkun Wang, Zhichao Lu, Qingfu Zhang
EM-MIAs: Enhancing Membership Inference Attacks in Large Language Models through Ensemble Modeling
Zichen Song, Sitan Huang, Zhongfeng Kang
Highly Optimized Kernels and Fine-Grained Codebooks for LLM Inference on Arm CPUs
Dibakar Gope, David Mansell, Danny Loh, Ian Bratt
On the Generalization Ability of Machine-Generated Text Detectors
Yule Liu, Zhiyuan Zhong, Yifan Liao, Zhen Sun, Jingyi Zheng, Jiaheng Wei, Qingyuan Gong, Fenghua Tong, Yang Chen, Yang Zhang, Xinlei He
Better Think with Tables: Leveraging Tables to Enhance Large Language Model Comprehension
Jio Oh, Geon Heo, Seungjun Oh, Jindong Wang, Xing Xie, Steven Euijong Whang
Hate Speech Detection and Target Identification in Devanagari Languages via Parameter Efficient Fine-Tuning of LLMs
Rushendra Sidibomma, Pransh Patwa, Parth Patwa, Aman Chadha, Vinija Jain, Amitava Das
DreamOmni: Unified Image Generation and Editing
Bin Xia, Yuechen Zhang, Jingyao Li, Chengyao Wang, Yitong Wang, Xinglong Wu, Bei Yu, Jiaya Jia
Analysis on LLMs Performance for Code Summarization
Md. Ahnaf Akib, Md. Muktadir Mazumder, Salman Ahsan
The HalluRAG Dataset: Detecting Closed-Domain Hallucinations in RAG Applications Using an LLM's Internal States
Fabian Ridder, Malte Schilling
DR-Encoder: Encode Low-rank Gradients with Random Prior for Large Language Models Differentially Privately
Huiwen Wu, Deyi Zhang, Xiaohan Li, Xiaogang Xu, Jiafei Wu, Zhe Liu
MINTQA: A Multi-Hop Question Answering Benchmark for Evaluating LLMs on New and Tail Knowledge
Jie He, Nan Hu, Wanqiu Long, Jiaoyan Chen, Jeff Z. Pan
Online Preference-based Reinforcement Learning with Self-augmented Feedback from Large Language Model
Songjun Tu, Jingbo Sun, Qichao Zhang, Xiangyuan Lan, Dongbin Zhao
Teaching LLMs to Refine with Tools
Dian Yu, Yuheng Zhang, Jiahao Xu, Tian Liang, Linfeng Song, Zhaopeng Tu, Haitao Mi, Dong Yu
Ask-Before-Detection: Identifying and Mitigating Conformity Bias in LLM-Powered Error Detector for Math Word Problem Solutions
Hang Li, Tianlong Xu, Kaiqi Yang, Yucheng Chu, Yanling Chen, Yichi Song, Qingsong Wen, Hui Liu
Argumentation Computation with Large Language Models : A Benchmark Study
Zhaoqun Li, Xiaotong Fang, Chen Chen, Mengze Li, Beishui Liao
Lillama: Large Language Models Compression via Low-Rank Feature Distillation
Yaya Sy, Christophe Cerisara, Irina Illina
The Task Shield: Enforcing Task Alignment to Defend Against Indirect Prompt Injection in LLM Agents
Feiran Jia, Tong Wu, Xin Qin, Anna Squicciarini
Internalized Self-Correction for Large Language Models
Nishanth Upadhyaya, Raghavendra Sridharamurthy
TimeRAG: BOOSTING LLM Time Series Forecasting via Retrieval-Augmented Generation
Silin Yang, Dong Wang, Haoqi Zheng, Ruochun Jin
Large Language Model Can Be a Foundation for Hidden Rationale-Based Retrieval
Luo Ji, Feixiang Guo, Teng Chen, Qingqing Gu, Xiaoyu Wang, Ningyuan Xi, Yihong Wang, Peng Yu, Yue Zhao, Hongyang Lei, Zhonglin Jiang, Yong Chen