Full Model
"Full Model" research encompasses the development and improvement of large-scale machine learning models across diverse applications, aiming to enhance performance, efficiency, and robustness. Current research focuses on addressing model vulnerabilities (e.g., adversarial attacks, hallucinations), improving efficiency for resource-constrained devices, and developing specialized models for specific domains (e.g., finance, astronomy, medical imaging). This work is significant for advancing AI capabilities in various fields and for mitigating potential risks associated with deploying complex models in real-world settings.
Papers
Evaluating the Robustness of Analogical Reasoning in Large Language Models
Martha Lewis, Melanie Mitchell
On the Fairness, Diversity and Reliability of Text-to-Image Generative Models
Jordan Vice, Naveed Akhtar, Richard Hartley, Ajmal Mian
Interactive and Expressive Code-Augmented Planning with Large Language Models
Anthony Z. Liu, Xinhe Wang, Jacob Sansom, Yao Fu, Jongwook Choi, Sungryull Sohn, Jaekyeom Kim, Honglak Lee
Tiny-Align: Bridging Automatic Speech Recognition and Large Language Model on the Edge
Ruiyang Qin, Dancheng Liu, Gelei Xu, Zheyu Yan, Chenhui Xu, Yuting Hu, X. Sharon Hu, Jinjun Xiong, Yiyu Shi
Benchmarking a wide range of optimisers for solving the Fermi-Hubbard model using the variational quantum eigensolver
Benjamin D.M. Jones, Lana Mineh, Ashley Montanaro
VideoAutoArena: An Automated Arena for Evaluating Large Multimodal Models in Video Analysis through User Simulation
Ziyang Luo, Haoning Wu, Dongxu Li, Jing Ma, Mohan Kankanhalli, Junnan Li
Puppet-CNN: Input-Adaptive Convolutional Neural Networks with Model Compression using Ordinary Differential Equation
Yucheng Xing, Xin Wang
Residual Vision Transformer (ResViT) Based Self-Supervised Learning Model for Brain Tumor Classification
Meryem Altin Karagoz, O. Ufuk Nalbantoglu, Geoffrey C. Fox
ACING: Actor-Critic for Instruction Learning in Black-Box Large Language Models
Salma Kharrat, Fares Fourati, Marco Canini
CodeXEmbed: A Generalist Embedding Model Family for Multiligual and Multi-task Code Retrieval
Ye Liu, Rui Meng, Shafiq Jot, Silvio Savarese, Caiming Xiong, Yingbo Zhou, Semih Yavuz
DLBacktrace: A Model Agnostic Explainability for any Deep Learning Models
Vinay Kumar Sankarapu, Chintan Chitroda, Yashwardhan Rathore, Neeraj Kumar Singh, Pratinav Seth
Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models
Laura Ruis, Maximilian Mozes, Juhan Bae, Siddhartha Rao Kamalakara, Dwarak Talupuru, Acyr Locatelli, Robert Kirk, Tim Rocktäschel, Edward Grefenstette, Max Bartolo
Frequency-Aware Guidance for Blind Image Restoration via Diffusion Models
Jun Xiao, Zihang Lyu, Hao Xie, Cong Zhang, Yakun Ju, Changjian Shui, Kin-Man Lam
JuniperLiu at CoMeDi Shared Task: Models as Annotators in Lexical Semantics Disagreements
Zhu Liu, Zhen Hu, Ying Liu
Benchmarking pre-trained text embedding models in aligning built asset information
Mehrzad Shahinmoghadam, Ali Motamedi
CNMBert: A Model for Hanyu Pinyin Abbreviation to Character Conversion Task
Zishuo Feng, Feng Cao
Re-examining learning linear functions in context
Omar Naim, Guilhem Fouilhé, Nicholas Asher
Exploring the Adversarial Vulnerabilities of Vision-Language-Action Models in Robotics
Taowen Wang, Dongfang Liu, James Chenhao Liang, Wenhao Yang, Qifan Wang, Cheng Han, Jiebo Luo, Ruixiang Tang