Full Model
"Full Model" research encompasses the development and improvement of large-scale machine learning models across diverse applications, aiming to enhance performance, efficiency, and robustness. Current research focuses on addressing model vulnerabilities (e.g., adversarial attacks, hallucinations), improving efficiency for resource-constrained devices, and developing specialized models for specific domains (e.g., finance, astronomy, medical imaging). This work is significant for advancing AI capabilities in various fields and for mitigating potential risks associated with deploying complex models in real-world settings.
Papers
Model-Based Differentially Private Knowledge Transfer for Large Language Models
Zhaomin Wu, Jizhou Guo, Junyi Hou, Bingsheng He, Lixin Fan, Qiang Yang
How to Leverage Demonstration Data in Alignment for Large Language Model? A Self-Imitation Learning Perspective
Teng Xiao, Mingxiao Li, Yige Yuan, Huaisheng Zhu, Chao Cui, Vasant G Honavar
Self-Data Distillation for Recovering Quality in Pruned Large Language Models
Vithursan Thangarasa, Ganesh Venkatesh, Nish Sinnadurai, Sean Lie
Dynamic Estimation of Learning Rates Using a Non-Linear Autoregressive Model
Ramin Okhrati
EBDM: Exemplar-guided Image Translation with Brownian-bridge Diffusion Models
Eungbean Lee, Somi Jeong, Kwanghoon Sohn
AM-SAM: Automated Prompting and Mask Calibration for Segment Anything Model
Yuchen Li, Li Zhang, Youwei Liang, Pengtao Xie
Transformers as Game Players: Provable In-context Game-playing Capabilities of Pre-trained Models
Chengshuai Shi, Kun Yang, Jing Yang, Cong Shen
Breaking the Memory Wall for Heterogeneous Federated Learning via Model Splitting
Chunlin Tian, Li Li, Kahou Tam, Yebo Wu, Chengzhong Xu
ControLRM: Fast and Controllable 3D Generation via Large Reconstruction Model
Hongbin Xu, Weitao Chen, Zhipeng Zhou, Feng Xiao, Baigui Sun, Mike Zheng Shou, Wenxiong Kang
Exploring Channel Distinguishability in Local Neighborhoods of the Model Space in Quantum Neural Networks
Sabrina Herbst, Sandeep Suresh Cranganore, Vincenzo De Maio, Ivona Brandic
Reinforcement Learning in Hyperbolic Spaces: Models and Experiments
Vladimir Jaćimović, Zinaid Kapić, Aladin Crnkić
Exact Aggregation for Federated and Efficient Fine-Tuning of Foundation Models
Raghav Singhal, Kaustubh Ponkshe, Praneeth Vepakomma
MITA: Bridging the Gap between Model and Data for Test-time Adaptation
Yige Yuan, Bingbing Xu, Teng Xiao, Liang Hou, Fei Sun, Huawei Shen, Xueqi Cheng
Mamba4Cast: Efficient Zero-Shot Time Series Forecasting with State Space Models
Sathya Kamesh Bhethanabhotla, Omar Swelam, Julien Siems, David Salinas, Frank Hutter
DARE the Extreme: Revisiting Delta-Parameter Pruning For Fine-Tuned Models
Wenlong Deng, Yize Zhao, Vala Vakilian, Minghui Chen, Xiaoxiao Li, Christos Thrampoulidis
\llinstruct: An Instruction-tuned model for English Language Proficiency Assessments
Debanjan Ghosh, Sophia Chan
EasyHeC++: Fully Automatic Hand-Eye Calibration with Pretrained Image Models
Zhengdong Hong, Kangfu Zheng, Linghao Chen
The Impact of Visual Information in Chinese Characters: Evaluating Large Models' Ability to Recognize and Utilize Radicals
Xiaofeng Wu, Karl Stratos, Wei Xu
Fragile Giants: Understanding the Susceptibility of Models to Subpopulation Attacks
Isha Gupta, Hidde Lycklama, Emanuel Opel, Evan Rose, Anwar Hithnawi
Context-Aware Full Body Anonymization using Text-to-Image Diffusion Models
Pascl Zwick, Kevin Roesch, Marvin Klemp, Oliver Bringmann