Full Model
"Full Model" research encompasses the development and improvement of large-scale machine learning models across diverse applications, aiming to enhance performance, efficiency, and robustness. Current research focuses on addressing model vulnerabilities (e.g., adversarial attacks, hallucinations), improving efficiency for resource-constrained devices, and developing specialized models for specific domains (e.g., finance, astronomy, medical imaging). This work is significant for advancing AI capabilities in various fields and for mitigating potential risks associated with deploying complex models in real-world settings.
Papers
SVDQunat: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models
Muyang Li, Yujun Lin, Zhekai Zhang, Tianle Cai, Xiuyu Li, Junxian Guo, Enze Xie, Chenlin Meng, Jun-Yan Zhu, Song Han
Diff-2-in-1: Bridging Generation and Dense Perception with Diffusion Models
Shuhong Zheng, Zhipeng Bao, Ruoyu Zhao, Martial Hebert, Yu-Xiong Wang
DISCO: DISCovering Overfittings as Causal Rules for Text Classification Models
Zijian Zhang, Vinay Setty, Yumeng Wang, Avishek Anand
TAP-VL: Text Layout-Aware Pre-training for Enriched Vision-Language Models
Jonathan Fhima, Elad Ben Avraham, Oren Nuriel, Yair Kittenplon, Roy Ganz, Aviad Aberdam, Ron Litman
One fish, two fish, but not the whole sea: Alignment reduces language models' conceptual diversity
Sonia K. Murthy, Tomer Ullman, Jennifer Hu
Repairing Neural Networks for Safety in Robotic Systems using Predictive Models
Keyvan Majd, Geoffrey Clark, Georgios Fainekos, Heni Ben Amor
Model and Deep learning based Dynamic Range Compression Inversion
Haoran Sun, Dominique Fourer, Hichem Maaref
Are Deep Learning Methods Suitable for Downscaling Global Climate Projections? Review and Intercomparison of Existing Models
Jose González-Abad, José Manuel Gutiérrez
Can Custom Models Learn In-Context? An Exploration of Hybrid Architecture Performance on In-Context Learning Tasks
Ryan Campbell, Nelson Lojo, Kesava Viswanadha, Christoffer Grondal Tryggestad, Derrick Han Sun, Sriteja Vijapurapu, August Rolfsen, Anant Sahai
A Novel Access Control and Privacy-Enhancing Approach for Models in Edge Computing
Peihao Li
Deferred Poisoning: Making the Model More Vulnerable via Hessian Singularization
Yuhao He, Jinyu Tian, Xianwei Zheng, Li Dong, Yuanman Li, Leo Yu Zhang, Jiantao Zhou
Zero-shot Dynamic MRI Reconstruction with Global-to-local Diffusion Model
Yu Guan, Kunlong Zhang, Qi Qi, Dong Wang, Ziwen Ke, Shaoyu Wang, Dong Liang, Qiegen Liu
DiffLM: Controllable Synthetic Data Generation via Diffusion Language Models
Ying Zhou, Xinyao Wang, Yulei Niu, Yaojie Shen, Lexin Tang, Fan Chen, Ben He, Le Sun, Longyin Wen
Advancing Recycling Efficiency: A Comparative Analysis of Deep Learning Models in Waste Classification
Zhanshan Qiao
Label Critic: Design Data Before Models
Pedro R. A. S. Bassi, Qilong Wu, Wenxuan Li, Sergio Decherchi, Andrea Cavalli, Alan Yuille, Zongwei Zhou
Multi-Transmotion: Pre-trained Model for Human Motion Prediction
Yang Gao, Po-Chien Luan, Alexandre Alahi
Continual LLaVA: Continual Instruction Tuning in Large Vision-Language Models
Meng Cao, Yuyang Liu, Yingfei Liu, Tiancai Wang, Jiahua Dong, Henghui Ding, Xiangyu Zhang, Ian Reid, Xiaodan Liang
Discrete the solving model of time-variant standard Sylvester-conjugate matrix equations using Euler-forward formula
Jiakuang He, Dongqing Wu
Evaluating Creative Short Story Generation in Humans and Large Language Models
Mete Ismayilzada, Claire Stevenson, Lonneke van der Plas