Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Thinking Forward and Backward: Effective Backward Planning with Large Language Models
Allen Z. Ren, Brian Ichter, Anirudha Majumdar
Towards Pedagogical LLMs with Supervised Fine Tuning for Computing Education
Alexandra Vassar, Jake Renzella, Emily Ross, Andrew Taylor
RAGViz: Diagnose and Visualize Retrieval-Augmented Generation
Tevin Wang, Jingyuan He, Chenyan Xiong
DynaSaur: Large Language Agents Beyond Predefined Actions
Dang Nguyen, Viet Dac Lai, Seunghyun Yoon, Ryan A. Rossi, Handong Zhao, Ruiyi Zhang, Puneet Mathur, Nedim Lipka, Yu Wang, Trung Bui, Franck Dernoncourt, Tianyi Zhou
Investigating Large Language Models for Complex Word Identification in Multilingual and Multidomain Setups
Răzvan-Alexandru Smădu, David-Gabriel Ion, Dumitru-Clementin Cercel, Florin Pop, Mihaela-Claudia Cercel
Data Extraction Attacks in Retrieval-Augmented Generation via Backdoors
Yuefeng Peng, Junda Wang, Hong Yu, Amir Houmansadr
An Exploration of Higher Education Course Evaluation by Large Language Models
Bo Yuan, Jiazi Hu
Graph-based Confidence Calibration for Large Language Models
Yukun Li, Sijia Wang, Lifu Huang, Li-Ping Liu
Ontology Population using LLMs
Sanaz Saki Norouzi, Adrita Barua, Antrea Christou, Nikita Gautam, Andrew Eells, Pascal Hitzler, Cogan Shimizu
Large Language Model Supply Chain: Open Problems From the Security Perspective
Qiang Hu, Xiaofei Xie, Sen Chen, Lei Ma
Are LLMs good pragmatic speakers?
Mingyue Jian, Siddharth Narayanaswamy
Enhancing LLM Evaluations: The Garbling Trick
William F. Bradley
High-performance automated abstract screening with large language model ensembles
Rohan Sanghera, Arun James Thirunavukarasu, Marc El Khoury, Jessica O'Logbon, Yuqing Chen, Archie Watt, Mustafa Mahmood, Hamid Butt, George Nishimura, Andrew Soltan
Sample-Efficient Alignment for LLMs
Zichen Liu, Changyu Chen, Chao Du, Wee Sun Lee, Min Lin
EEE-Bench: A Comprehensive Multimodal Electrical And Electronics Engineering Benchmark
Ming Li, Jike Zhong, Tianle Chen, Yuxiang Lai, Konstantinos Psounis
A Deep Dive Into Large Language Model Code Generation Mistakes: What and Why?
QiHong Chen, Jiawei Li, Jiecheng Deng, Jiachen Yu, Justin Tian Jin Chen, Iftekhar Ahmed
Interacting Large Language Model Agents. Interpretable Models and Social Learning
Adit Jain, Vikram Krishnamurthy
Diversidade linguística e inclusão digital: desafios para uma ia brasileira
Raquel Meister Ko Freitag
PMoL: Parameter Efficient MoE for Preference Mixing of LLM Alignment
Dongxu Liu, Bing Xu, Yinzhuo Chen, Bufan Xu, Wenpeng Lu, Muyun Yang, Tiejun Zhao
Transfer Learning for Finetuning Large Language Models
Tobias Strangmann, Lennart Purucker, Jörg K.H. Franke, Ivo Rapant, Fabio Ferreira, Frank Hutter