Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Controlling the Mutation in Large Language Models for the Efficient Evolution of Algorithms
Haoran Yin, Anna V. Kononova, Thomas Bäck, Niki van Stein
AIM: Adaptive Inference of Multi-Modal LLMs via Token Merging and Pruning
Yiwu Zhong, Zhuoming Liu, Yin Li, Liwei Wang
Does Safety Training of LLMs Generalize to Semantically Related Natural Prompts?
Sravanti Addepalli, Yerram Varun, Arun Suggala, Karthikeyan Shanmugam, Prateek Jain
Linq-Embed-Mistral Technical Report
Chanyeol Choi, Junseong Kim, Seolhwa Lee, Jihoon Kwon, Sangmo Gu, Yejin Kim, Minkyung Cho, Jy-yong Sohn
Automatic detection of diseases in Spanish clinical notes combining medical language models and ontologies
Leon-Paul Schaub Torre, Pelayo Quiros, Helena Garcia Mieres
Byte BPE Tokenization as an Inverse string Homomorphism
Saibo Geng, Sankalp Gambhir, Chris Wendler, Robert West
Fine-Grained Behavior Simulation with Role-Playing Large Language Model on Social Media
Kun Li, Chenwei Dai, Wei Zhou, Songlin Hu
Unifying KV Cache Compression for Large Language Models with LeanKV
Yanqi Zhang, Yuwei Hu, Runyuan Zhao, John C.S. Lui, Haibo Chen
ChatTS: Aligning Time Series with LLMs via Synthetic Data for Enhanced Understanding and Reasoning
Zhe Xie, Zeyan Li, Xiao He, Longlong Xu, Xidao Wen, Tieying Zhang, Jianjun Chen, Rui Shi, Dan Pei
Revolve: Optimizing AI Systems by Tracking Response Evolution in Textual Optimization
Peiyan Zhang, Haibo Jin, Leyang Hu, Xinnuo Li, Liying Kang, Man Luo, Yangqiu Song, Haohan Wang
CBEval: A framework for evaluating and interpreting cognitive biases in LLMs
Ammar Shaikh, Raj Abhijit Dandekar, Sreedath Panat, Rajat Dandekar
Human Variability vs. Machine Consistency: A Linguistic Analysis of Texts Generated by Humans and Large Language Models
Sergio E. Zanotto, Segun Aroyehun
Curriculum-style Data Augmentation for LLM-based Metaphor Detection
Kaidi Jia, Yanxia Wu, Rongsheng Li
Enhancing Trust in Large Language Models with Uncertainty-Aware Fine-Tuning
Ranganath Krishnan, Piyush Khanna, Omesh Tickoo
Removing Spurious Correlation from Neural Network Interpretations
Milad Fotouhi, Mohammad Taha Bahadori, Oluwaseyi Feyisetan, Payman Arabshahi, David Heckerman
Measuring Bias of Web-filtered Text Datasets and Bias Propagation Through Training
Youssef Mansour, Reinhard Heckel
Flattering to Deceive: The Impact of Sycophantic Behavior on User Trust in Large Language Model
María Victoria Carro
An Evolutionary Large Language Model for Hallucination Mitigation
Abdennour Boulesnane, Abdelhakim Souilah
Optimizing Large Language Models for Turkish: New Methodologies in Corpus Selection and Training
H. Toprak Kesgin, M. Kaan Yuce, Eren Dogan, M. Egemen Uzun, Atahan Uz, Elif Ince, Yusuf Erdem, Osama Shbib, Ahmed Zeer, M. Fatih Amasyali
Cosmos-LLaVA: Chatting with the Visual Cosmos-LLaVA: Görselle Sohbet Etmek
Ahmed Zeer, Eren Dogan, Yusuf Erdem, Elif Ince, Osama Shbib, M. Egemen Uzun, Atahan Uz, M. Kaan Yuce, H. Toprak Kesgin, M. Fatih Amasyali