Full Model
"Full Model" research encompasses the development and improvement of large-scale machine learning models across diverse applications, aiming to enhance performance, efficiency, and robustness. Current research focuses on addressing model vulnerabilities (e.g., adversarial attacks, hallucinations), improving efficiency for resource-constrained devices, and developing specialized models for specific domains (e.g., finance, astronomy, medical imaging). This work is significant for advancing AI capabilities in various fields and for mitigating potential risks associated with deploying complex models in real-world settings.
Papers
NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts
Yen-Ting Lin, Chao-Han Huck Yang, Zhehuai Chen, Piotr Zelasko, Xuesong Yang, Zih-Ching Chen, Krishna C Puvvada, Szu-Wei Fu, Ke Hu, Jun Wei Chiu, Jagadeesh Balam, Boris Ginsburg, Yu-Chiang Frank Wang
The effect of different feature selection methods on models created with XGBoost
Jorge Neyra, Vishal B. Siramshetty, Huthaifa I. Ashqar
LLMs as Method Actors: A Model for Prompt Engineering and Architecture
Colin Doyle
HeartBERT: A Self-Supervised ECG Embedding Model for Efficient and Effective Medical Signal Analysis
Saedeh Tahery, Fatemeh Hamid Akhlaghi, Termeh Amirsoleimani, Saeed Farzi, Carlo Strapparava
SVDQunat: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models
Muyang Li, Yujun Lin, Zhekai Zhang, Tianle Cai, Xiuyu Li, Junxian Guo, Enze Xie, Chenlin Meng, Jun-Yan Zhu, Song Han
Diff-2-in-1: Bridging Generation and Dense Perception with Diffusion Models
Shuhong Zheng, Zhipeng Bao, Ruoyu Zhao, Martial Hebert, Yu-Xiong Wang
DISCO: DISCovering Overfittings as Causal Rules for Text Classification Models
Zijian Zhang, Vinay Setty, Yumeng Wang, Avishek Anand
TAP-VL: Text Layout-Aware Pre-training for Enriched Vision-Language Models
Jonathan Fhima, Elad Ben Avraham, Oren Nuriel, Yair Kittenplon, Roy Ganz, Aviad Aberdam, Ron Litman
One fish, two fish, but not the whole sea: Alignment reduces language models' conceptual diversity
Sonia K. Murthy, Tomer Ullman, Jennifer Hu
Repairing Neural Networks for Safety in Robotic Systems using Predictive Models
Keyvan Majd, Geoffrey Clark, Georgios Fainekos, Heni Ben Amor
Model and Deep learning based Dynamic Range Compression Inversion
Haoran Sun, Dominique Fourer, Hichem Maaref
Are Deep Learning Methods Suitable for Downscaling Global Climate Projections? Review and Intercomparison of Existing Models
Jose González-Abad, José Manuel Gutiérrez
Can Custom Models Learn In-Context? An Exploration of Hybrid Architecture Performance on In-Context Learning Tasks
Ryan Campbell, Nelson Lojo, Kesava Viswanadha, Christoffer Grondal Tryggestad, Derrick Han Sun, Sriteja Vijapurapu, August Rolfsen, Anant Sahai
A Novel Access Control and Privacy-Enhancing Approach for Models in Edge Computing
Peihao Li
Deferred Poisoning: Making the Model More Vulnerable via Hessian Singularization
Yuhao He, Jinyu Tian, Xianwei Zheng, Li Dong, Yuanman Li, Jiantao Zhou
Zero-shot Dynamic MRI Reconstruction with Global-to-local Diffusion Model
Yu Guan, Kunlong Zhang, Qi Qi, Dong Wang, Ziwen Ke, Shaoyu Wang, Dong Liang, Qiegen Liu
DiffLM: Controllable Synthetic Data Generation via Diffusion Language Models
Ying Zhou, Xinyao Wang, Yulei Niu, Yaojie Shen, Lexin Tang, Fan Chen, Ben He, Le Sun, Longyin Wen
Advancing Recycling Efficiency: A Comparative Analysis of Deep Learning Models in Waste Classification
Zhanshan Qiao
Label Critic: Design Data Before Models
Pedro R. A. S. Bassi, Qilong Wu, Wenxuan Li, Sergio Decherchi, Andrea Cavalli, Alan Yuille, Zongwei Zhou