Full Model
"Full Model" research encompasses the development and improvement of large-scale machine learning models across diverse applications, aiming to enhance performance, efficiency, and robustness. Current research focuses on addressing model vulnerabilities (e.g., adversarial attacks, hallucinations), improving efficiency for resource-constrained devices, and developing specialized models for specific domains (e.g., finance, astronomy, medical imaging). This work is significant for advancing AI capabilities in various fields and for mitigating potential risks associated with deploying complex models in real-world settings.
Papers
ACING: Actor-Critic for Instruction Learning in Black-Box Large Language Models
Salma Kharrat, Fares Fourati, Marco Canini
CodeXEmbed: A Generalist Embedding Model Family for Multiligual and Multi-task Code Retrieval
Ye Liu, Rui Meng, Shafiq Jot, Silvio Savarese, Caiming Xiong, Yingbo Zhou, Semih Yavuz
DLBacktrace: A Model Agnostic Explainability for any Deep Learning Models
Vinay Kumar Sankarapu, Chintan Chitroda, Yashwardhan Rathore, Neeraj Kumar Singh, Pratinav Seth
Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models
Laura Ruis, Maximilian Mozes, Juhan Bae, Siddhartha Rao Kamalakara, Dwarak Talupuru, Acyr Locatelli, Robert Kirk, Tim Rocktäschel, Edward Grefenstette, Max Bartolo
Frequency-Aware Guidance for Blind Image Restoration via Diffusion Models
Jun Xiao, Zihang Lyu, Hao Xie, Cong Zhang, Yakun Ju, Changjian Shui, Kin-Man Lam
CoMeDi Shared Task: Models as Annotators in Lexical Semantics Disagreements
Zhu Liu, Zhen Hu, Ying Liu
Benchmarking pre-trained text embedding models in aligning built asset information
Mehrzad Shahinmoghadam, Ali Motamedi
CNMBert: A Model For Hanyu Pinyin Abbreviation to Character Conversion Task
Zishuo Feng, Feng Cao
Re-examining learning linear functions in context
Omar Naim, Guilhem Fouilhé, Nicholas Asher
MARS: Unleashing the Power of Variance Reduction for Training Large Models
Huizhuo Yuan, Yifeng Liu, Shuang Wu, Xun Zhou, Quanquan Gu
OnlyFlow: Optical Flow based Motion Conditioning for Video Diffusion Models
Mathis Koroglu, Hugo Caselles-Dupré, Guillaume Jeanneret Sanmiguel, Matthieu Cord
Explanation for Trajectory Planning using Multi-modal Large Language Model for Autonomous Driving
Shota Yamazaki, Chenyu Zhang, Takuya Nanri, Akio Shigekane, Siyuan Wang, Jo Nishiyama, Tao Chu, Kohei Yokosawa
DT-JRD: Deep Transformer based Just Recognizable Difference Prediction Model for Video Coding for Machines
Junqi Liu, Yun Zhang, Xiaoqi Wang, Xu Long, Sam Kwong
Enhancing Financial Domain Adaptation of Language Models via Model Augmentation
Kota Tanabe, Masanori Hirano, Kazuki Matoya, Kentaro Imajo, Hiroki Sakaji, Itsuki Noda
Model agnostic local variable importance for locally dependent relationships
Kelvyn K. Bladen, Adele Cutler, D. Richard Cutler, Kevin R. Moon
Towards Optimizing a Retrieval Augmented Generation using Large Language Model on Academic Data
Anum Afzal, Juraj Vladika, Gentrit Fazlija, Andrei Staradubets, Florian Matthes
Deceiving Question-Answering Models: A Hybrid Word-Level Adversarial Approach
Jiyao Li, Mingze Ni, Yongshun Gong, Wei Liu
PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Model
Yilun Liu, Yunpu Ma, Shuo Chen, Zifeng Ding, Bailan He, Zhen Han, Volker Tresp
LLMPhy: Complex Physical Reasoning Using Large Language Models and World Models
Anoop Cherian, Radu Corcodel, Siddarth Jain, Diego Romeres