Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Predicting the Performance of Black-box LLMs through Self-Queries
Dylan Sam, Marc Finzi, J. Zico Kolter
Many of Your DPOs are Secretly One: Attempting Unification Through Mutual Information
Rasul Tutnov, Antoine Grosnit, Haitham Bou-Ammar
Decoding Knowledge in Large Language Models: A Framework for Categorization and Comprehension
Yanbo Fang, Ruixiang Tang
Think More, Hallucinate Less: Mitigating Hallucinations via Dual Process of Fast and Slow Thinking
Xiaoxue Cheng, Junyi Li, Wayne Xin Zhao, Ji-Rong Wen
Large Language Models for Mental Health Diagnostic Assessments: Exploring The Potential of Large Language Models for Assisting with Mental Health Diagnostic Assessments -- The Depression and Anxiety Case
Kaushik Roy, Harshul Surana, Darssan Eswaramoorthi, Yuxin Zi, Vedant Palit, Ritvik Garimella, Amit Sheth
Does a Large Language Model Really Speak in Human-Like Language?
Mose Park, Yunjin Choi, Jong-June Jeon
Digital Guardians: Can GPT-4, Perspective API, and Moderation API reliably detect hate speech in reader comments of German online newspapers?
Manuel Weber, Moritz Huber, Maximilian Auch, Alexander Döschl, Max-Emanuel Keller, Peter Mandl
Large Language Model-Enhanced Symbolic Reasoning for Knowledge Base Completion
Qiyuan He, Jianfei Yu, Wenya Wang
Enhancing Reasoning through Process Supervision with Monte Carlo Tree Search
Shuangtao Li, Shuaihao Dong, Kexin Luan, Xinhan Di, Chaofan Ding
Harnessing Multi-Agent LLMs for Complex Engineering Problem-Solving: A Framework for Senior Design Projects
Abdullah Mushtaq, Muhammad Rafay Naeem, Ibrahim Ghaznavi, Muhammad Imran Taj, Imran Hashmi, Junaid Qadir
Dynamic Attention-Guided Context Decoding for Mitigating Context Faithfulness Hallucinations in Large Language Models
Yanwen Huang, Yong Zhang, Ning Cheng, Zhitao Li, Shaojun Wang, Jing Xiao
KaLM-Embedding: Superior Training Data Brings A Stronger Embedding Model
Xinshuo Hu, Zifei Shan, Xinping Zhao, Zetian Sun, Zhenyu Liu, Dongfang Li, Shaolin Ye, Xinyuan Wei, Qian Chen, Baotian Hu, Haofen Wang, Jun Yu, Min Zhang
MDSF: Context-Aware Multi-Dimensional Data Storytelling Framework based on Large language Model
Chengze Zhang, Changshan Li, Shiyang Gao
Aligning LLMs with Domain Invariant Reward Models
David Wu, Sanjiban Choudhury
Large Language Model Based Multi-Agent System Augmented Complex Event Processing Pipeline for Internet of Multimedia Things
Talha Zeeshan, Abhishek Kumar, Susanna Pirttikangas, Sasu Tarkoma
Representation in large language models
Cameron C. Yetman
TrustRAG: Enhancing Robustness and Trustworthiness in RAG
Huichi Zhou, Kin-Hei Lee, Zhonghao Zhan, Yue Chen, Zhenhao Li
LUSIFER: Language Universal Space Integration for Enhanced Multilingual Embeddings with Large Language Models
Hieu Man, Nghia Trung Ngo, Viet Dac Lai, Ryan A. Rossi, Franck Dernoncourt, Thien Huu Nguyen
NMM-HRI: Natural Multi-modal Human-Robot Interaction with Voice and Deictic Posture via Large Language Model
Yuzhi Lai, Shenghai Yuan, Youssef Nassar, Mingyu Fan, Atmaraaj Gopal, Arihiro Yorita, Naoyuki Kubota, Matthias Rätsch