Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Textualized Agent-Style Reasoning for Complex Tasks by Multiple Round LLM Generation
Chen Liang, Zhifan Feng, Zihe Liu, Wenbin Jiang, Jinan Xu, Yufeng Chen, Yong Wang
On the Effectiveness of LLMs for Manual Test Verifications
Myron David Lucena Campos Peixoto, Davy de Medeiros Baia, Nathalia Nascimento, Paulo Alencar, Baldoino Fonseca, Márcio Ribeiro
RAG-Modulo: Solving Sequential Tasks using Experience, Critics, and Language Models
Abhinav Jain, Chris Jermaine, Vaibhav Unhelkar
Bootstrapping Object-level Planning with Large Language Models
David Paulius, Alejandro Agostini, Benedict Quartey, George Konidaris
Fine-Tuning a Time Series Foundation Model with Wasserstein Loss
Andrei Chernov
Gender Representation and Bias in Indian Civil Service Mock Interviews
Somonnoy Banerjee, Sujan Dutta, Soumyajit Datta, Ashiqur R. KhudaBukhsh
To CoT or not to CoT? Chain-of-thought helps mainly on math and symbolic reasoning
Zayne Sprague, Fangcong Yin, Juan Diego Rodriguez, Dongwei Jiang, Manya Wadhwa, Prasann Singhal, Xinyu Zhao, Xi Ye, Kyle Mahowald, Greg Durrett
Finetuning Language Models to Emit Linguistic Expressions of Uncertainty
Arslan Chaudhry, Sridhar Thiagarajan, Dilan Gorur
Decoding Style: Efficient Fine-Tuning of LLMs for Image-Guided Outfit Recommendation with Preference
Najmeh Forouzandehmehr, Nima Farrokhsiar, Ramin Giahi, Evren Korpeoglu, Kannan Achan
BodyShapeGPT: SMPL Body Shape Manipulation with LLMs
Baldomero R. Árbol, Dan Casas
VERA: Validation and Enhancement for Retrieval Augmented systems
Nitin Aravind Birur, Tanay Baswa, Divyanshu Kumar, Jatan Loya, Sahil Agarwal, Prashanth Harshangi
Dual-Layer Training and Decoding of Large Language Model with Simultaneously Thinking and Speaking
Ningyuan Xi, Xiaoyu Wang, Yetao Wu, Teng Chen, Qingqing Gu, Jinxian Qu, Zhonglin Jiang, Yong Chen, Luo Ji
Using Large Language Models to Generate Clinical Trial Tables and Figures
Yumeng Yang, Peter Krusche, Kristyn Pantoja, Cheng Shi, Ethan Ludmir, Kirk Roberts, Gen Zhu
Sampling Latent Material-Property Information From LLM-Derived Embedding Representations
Luke P. J. Gilligan, Matteo Cobelli, Hasan M. Sayeed, Taylor D. Sparks, Stefano Sanvito
Efficacy of Synthetic Data as a Benchmark
Gaurav Maheshwari, Dmitry Ivanov, Kevin El Haddad
LLM-wrapper: Black-Box Semantic-Aware Adaptation of Vision-Language Models for Referring Expression Comprehension
Amaia Cardiel, Eloi Zablocki, Elias Ramzi, Oriane Siméoni, Matthieu Cord
LLMs in Education: Novel Perspectives, Challenges, and Opportunities
Bashar Alhafni, Sowmya Vajjala, Stefano Bannò, Kaushal Kumar Maurya, Ekaterina Kochmar
LLMs + Persona-Plug = Personalized LLMs
Jiongnan Liu, Yutao Zhu, Shuting Wang, Xiaochi Wei, Erxue Min, Yu Lu, Shuaiqiang Wang, Dawei Yin, Zhicheng Dou
Local Explanations and Self-Explanations for Assessing Faithfulness in black-box LLMs
Christos Fragkathoulas, Odysseas S. Chlapanis
The Factuality of Large Language Models in the Legal Domain
Rajaa El Hamdani, Thomas Bonald, Fragkiskos Malliaros, Nils Holzenberger, Fabian Suchanek