Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
DSAI: Unbiased and Interpretable Latent Feature Extraction for Data-Centric AI
Hyowon Cho, Soonwon Ka, Daechul Park, Jaewook Kang, Minjoon Seo, Bokyung Son
Methods for Legal Citation Prediction in the Age of LLMs: An Australian Law Case Study
Ehsan Shareghi, Jiuzhou Han, Paul Burgess
Blockchain Data Analysis in the Era of Large-Language Models
Kentaroh Toyoda, Xiao Wang, Mingzhe Li, Bo Gao, Yuan Wang, Qingsong Wei
Unseen Attack Detection in Software-Defined Networking Using a BERT-Based Large Language Model
Mohammed N. Swileh (1), Shengli Zhang (1) ((1) College of Electronics and Information Engineering, Shenzhen University, Shenzhen, China)
LLMs as Debate Partners: Utilizing Genetic Algorithms and Adversarial Search for Adaptive Arguments
Prakash Aryan
Enhancing Adversarial Resistance in LLMs with Recursion
Bryan Li, Sounak Bagchi, Zizhan Wang
Hate Speech According to the Law: An Analysis for Effective Detection
Katerina Korre, John Pavlopoulos, Paolo Gajo, Alberto Barrón-Cedeño
AIDE: Task-Specific Fine Tuning with Attribute Guided Multi-Hop Data Expansion
Jiayu Li, Xuan Zhu, Fang Liu, Yanjun Qi
Infusing Prompts with Syntax and Semantics
Anton Bulle Labate, Fabio Gagliardi Cozman
KaSA: Knowledge-Aware Singular-Value Adaptation of Large Language Models
Fan Wang, Juyong Jiang, Chansung Park, Sunghun Kim, Jing Tang
Evaluating Robustness of LLMs on Crisis-Related Microblogs across Events, Information Types, and Linguistic Features
Muhammad Imran, Abdul Wahab Ziaullah, Kai Chen, Ferda Ofli
Cooperative SQL Generation for Segmented Databases By Using Multi-functional LLM Agents
Zhiguang Wu, Fengbin Zhu, Xuequn Shang, Yupei Zhang, Pan Zhou
Are Clinical T5 Models Better for Clinical Text?
Yahan Li, Keith Harrigian, Ayah Zirikly, Mark Dredze
Training-Free Bayesianization for Low-Rank Adapters of Large Language Models
Haizhou Shi, Yibin Wang, Ligong Han, Huan Zhang, Hao Wang
Semantic loss guided data efficient supervised fine tuning for Safe Responses in LLMs
Yuxiao Lu, Arunesh Sinha, Pradeep Varakantham
CharacterBox: Evaluating the Role-Playing Capabilities of LLMs in Text-Based Virtual Worlds
Lei Wang, Jianxun Lian, Yi Huang, Yanqi Dai, Haoxuan Li, Xu Chen, Xing Xie, Ji-Rong Wen
Towards Learning to Reason: Comparing LLMs with Neuro-Symbolic on Arithmetic Relations in Abstract Reasoning
Michael Hersche, Giacomo Camposampiero, Roger Wattenhofer, Abu Sebastian, Abbas Rahimi
LLMs-as-Judges: A Comprehensive Survey on LLM-based Evaluation Methods
Haitao Li, Qian Dong, Junjie Chen, Huixue Su, Yujia Zhou, Qingyao Ai, Ziyi Ye, Yiqun Liu
A polar coordinate system represents syntax in large language models
Pablo Diego-Simón, Stéphane D'Ascoli, Emmanuel Chemla, Yair Lakretz, Jean-Rémi King
A Survey on Uncertainty Quantification of Large Language Models: Taxonomy, Open Research Challenges, and Future Directions
Ola Shorinwa, Zhiting Mei, Justin Lidard, Allen Z. Ren, Anirudha Majumdar