Large Model
Large models, encompassing large language models (LLMs) and their multimodal counterparts (MLLMs), are rapidly advancing artificial intelligence by leveraging massive datasets and immense computational power to achieve state-of-the-art performance across diverse tasks. Current research emphasizes efficient fine-tuning techniques, including model compression and low-rank adaptation, to address the challenges of deploying these resource-intensive models, as well as improving their robustness and uncertainty quantification. These advancements are driving progress in various fields, from improved search engines and medical image analysis to novel applications in robotics, finance, and agriculture. The development of robust evaluation benchmarks and the exploration of the interplay between large and small models are also key areas of focus.
Papers
MME-Finance: A Multimodal Finance Benchmark for Expert-level Understanding and Reasoning
Ziliang Gan, Yu Lu, Dong Zhang, Haohan Li, Che Liu, Jian Liu, Ji Liu, Haipang Wu, Chaoyou Fu, Zenglin Xu, Rongjunchen Zhang, Yong Dai
Inference Optimal VLMs Need Only One Visual Token but Larger Models
Kevin Y. Li, Sachin Goyal, Joao D. Semedo, J. Zico Kolter
Specialized Foundation Models Struggle to Beat Supervised Baselines
Zongzhe Xu, Ritvik Gupta, Wenduo Cheng, Alexander Shen, Junhong Shen, Ameet Talwalkar, Mikhail Khodak
Larger models yield better results? Streamlined severity classification of ADHD-related concerns using BERT-based knowledge distillation
Ahmed Akib Jawad Karim, Kazi Hafiz Md. Asad, Md. Golam Rabiul Alam
Keypoint Abstraction using Large Models for Object-Relative Imitation Learning
Xiaolin Fang, Bo-Ruei Huang, Jiayuan Mao, Jasmine Shone, Joshua B. Tenenbaum, Tomás Lozano-Pérez, Leslie Pack Kaelbling
A Comprehensive Study on Quantization Techniques for Large Language Models
Jiedong Lang, Zhehao Guo, Shuyu Huang
Large Legislative Models: Towards Efficient AI Policymaking in Economic Simulations
Henry Gasztowtt, Benjamin Smith, Vincent Zhu, Qinxun Bai, Edwin Zhang
Packing Analysis: Packing Is More Appropriate for Large Models or Datasets in Supervised Fine-tuning
Shuhe Wang, Guoyin Wang, Yizhong Wang, Jiwei Li, Eduard Hovy, Chen Guo