Large Language Model
Large language models (LLMs) are sophisticated AI systems designed to process and generate human-like text, aiming to improve various natural language processing tasks. Current research focuses on enhancing LLM safety, efficiency (through techniques like quantization and optimized decoding), and fairness, as well as improving their ability to perform complex reasoning and handle diverse instructions. These advancements are significant because they address critical limitations in current LLMs and pave the way for broader applications across diverse fields, including healthcare, legal tech, and autonomous systems.
Papers
Using off-the-shelf LLMs to query enterprise data by progressively revealing ontologies
C. Civili, E. Sherkhonov, R.E.K. Stirewalt
Enhancing Long Context Performance in LLMs Through Inner Loop Query Mechanism
Yimin Tang, Yurong Xu, Ning Yan, Masood Mortazavi
P-FOLIO: Evaluating and Improving Logical Reasoning with Abundant Human-Written Reasoning Chains
Simeng Han, Aaron Yu, Rui Shen, Zhenting Qi, Martin Riddell, Wenfei Zhou, Yujie Qiao, Yilun Zhao, Semih Yavuz, Ye Liu, Shafiq Joty, Yingbo Zhou, Caiming Xiong, Dragomir Radev, Rex Ying, Arman Cohan
Can a large language model be a gaslighter?
Wei Li, Luyao Zhu, Yang Song, Ruixi Lin, Rui Mao, Yang You
Enterprise Benchmarks for Large Language Model Evaluation
Bing Zhang, Mikio Takeuchi, Ryo Kawahara, Shubhi Asthana, Md. Maruf Hossain, Guang-Jie Ren, Kate Soule, Yada Zhu
Hybrid Training Approaches for LLMs: Leveraging Real and Synthetic Data to Enhance Model Performance in Domain-Specific Applications
Alexey Zhezherau, Alexei Yanockin
SimpleStrat: Diversifying Language Model Generation with Stratification
Justin Wong, Yury Orlovskiy, Michael Luo, Sanjit A. Seshia, Joseph E. Gonzalez
Optimized Biomedical Question-Answering Services with LLM and Multi-BERT Integration
Cheng Qian, Xianglong Shi, Shanshan Yao, Yichen Liu, Fengming Zhou, Zishu Zhang, Junaid Akram, Ali Braytee, Ali Anaissi
The structure of the token space for large language models
Michael Robinson, Sourya Dey, Shauna Sweet
SubZero: Random Subspace Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning
Ziming Yu, Pan Zhou, Sike Wang, Jia Li, Hua Huang
Controllable Safety Alignment: Inference-Time Adaptation to Diverse Safety Requirements
Jingyu Zhang, Ahmed Elgohary, Ahmed Magooda, Daniel Khashabi, Benjamin Van Durme
Maximizing the Potential of Synthetic Data: Insights from Random Matrix Theory
Aymane El Firdoussi, Mohamed El Amine Seddik, Soufiane Hayou, Reda Alami, Ahmed Alzubaidi, Hakim Hacid
A Benchmark for Cross-Domain Argumentative Stance Classification on Social Media
Jiaqing Yuan, Ruijie Xi, Munindar P. Singh
Audio Description Generation in the Era of LLMs and VLMs: A Review of Transferable Generative AI Technologies
Yingqiang Gao, Lukas Fischer, Alexa Lintner, Sarah Ebling
Measuring the Inconsistency of Large Language Models in Preferential Ranking
Xiutian Zhao, Ke Wang, Wei Peng
Do Unlearning Methods Remove Information from Language Model Weights?
Aghyad Deeb, Fabien Roger
StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Zhuoqun Li, Xuanang Chen, Haiyang Yu, Hongyu Lin, Yaojie Lu, Qiaoyu Tang, Fei Huang, Xianpei Han, Le Sun, Yongbin Li
Don't Transform the Code, Code the Transforms: Towards Precise Code Rewriting using LLMs
Chris Cummins, Volker Seeker, Jordi Armengol-Estapé, Aram H. Markosyan, Gabriel Synnaeve, Hugh Leather
Superpipeline: A Universal Approach for Reducing GPU Memory Usage in Large Models
Reza Abbasi, Sernam Lim
Developing a Pragmatic Benchmark for Assessing Korean Legal Language Understanding in Large Language Models
Yeeun Kim, Young Rok Choi, Eunkyung Choi, Jinhwan Choi, Hai Jin Park, Wonseok Hwang