Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Approximated Variational Bayesian Inverse Reinforcement Learning for Large Language Model Alignment
Yuang Cai, Yuyu Yuan, Jinsheng Shi, Qinhong Lin
StreamAdapter: Efficient Test Time Adaptation from Contextual Streams
Dilxat Muhtar, Yelong Shen, Yaming Yang, Xiaodong Liu, Yadong Lu, Jianfeng Liu, Yuefeng Zhan, Hao Sun, Weiwei Deng, Feng Sun, Xueliang Zhang, Jianfeng Gao, Weizhu Chen, Qi Zhang
Enhancing Financial Domain Adaptation of Language Models via Model Augmentation
Kota Tanabe, Masanori Hirano, Kazuki Matoya, Kentaro Imajo, Hiroki Sakaji, Itsuki Noda
HateGPT: Unleashing GPT-3.5 Turbo to Combat Hate Speech on X
Aniket Deroy, Subhankar Maity
Comprehensive and Practical Evaluation of Retrieval-Augmented Generation Systems for Medical Question Answering
Nghia Trung Ngo, Chien Van Nguyen, Franck Dernoncourt, Thien Huu Nguyen
P-MMEval: A Parallel Multilingual Multitask Benchmark for Consistent Evaluation of LLMs
Yidan Zhang, Boyi Deng, Yu Wan, Baosong Yang, Haoran Wei, Fei Huang, Bowen Yu, Junyang Lin, Fei Huang, Jingren Zhou
Reducing Reasoning Costs - The Path of Optimization for Chain of Thought via Sparse Attention Mechanism
Libo Wang
Code-mixed LLM: Improve Large Language Models' Capability to Handle Code-Mixing through Reinforcement Learning from AI Feedback
Wenbo Zhang, Aditya Majumdar, Amulya Yadav
Cut Your Losses in Large-Vocabulary Language Models
Erik Wijmans, Brody Huval, Alexander Hertzberg, Vladlen Koltun, Philipp Krähenbühl
Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection
Vima Gupta, Kartik Sinha, Ada Gavrilovska, Anand Padmanabha Iyer
CoCoP: Enhancing Text Classification with LLM through Code Completion Prompt
Mohammad Mahdi Mohajeri, Mohammad Javad Dousti, Majid Nili Ahmadabadi
The Limited Impact of Medical Adaptation of Large Language and Vision-Language Models
Daniel P. Jeong, Pranav Mani, Saurabh Garg, Zachary C. Lipton, Michael Oberst
XiYan-SQL: A Multi-Generator Ensemble Framework for Text-to-SQL
Yingqi Gao, Yifu Liu, Xiaoxia Li, Xiaorong Shi, Yin Zhu, Yiming Wang, Shiqi Li, Wei Li, Yuntao Hong, Zhiling Luo, Jinyang Gao, Liyu Mou, Yu Li
Leveraging LLMs for Predictive Insights in Food Policy and Behavioral Interventions
Micha Kaiser, Paul Lohmann, Peter Ochieng, Billy Shi, Cass R. Sunstein, Lucia A. Reisch
LogLLM: Log-based Anomaly Detection Using Large Language Models
Wei Guan, Jian Cao, Shiyou Qian, Jianqi Gao
CorrSynth -- A Correlated Sampling Method for Diverse Dataset Generation from LLMs
Suhas S Kowshik, Abhishek Divekar, Vijit Malik
One STEP at a time: Language Agents are Stepwise Planners
Minh Nguyen, Ehsan Shareghi
Quantifying Qualitative Insights: Leveraging LLMs to Market Predict
Hoyoung Lee, Youngsoo Choi, Yuhee Kwon
Refining Translations with LLMs: A Constraint-Aware Iterative Prompting Approach
Shangfeng Chen, Xiayang Shi, Pu Li, Yinlin Li, Jingjing Liu
R3HF: Reward Redistribution for Enhancing Reinforcement Learning from Human Feedback
Jiahui Li, Tai-wei Chang, Fengda Zhang, Kun Kuang, Long Chen