Deep Model
Deep models, encompassing a broad range of neural network architectures, aim to learn complex patterns from data for various tasks like image classification, time series forecasting, and system identification. Current research emphasizes improving efficiency (e.g., through constant-time learning algorithms and layer caching), enhancing explainability (e.g., via gradient-free methods), and mitigating issues like bias and memorization. These advancements are significant because they improve the reliability, trustworthiness, and applicability of deep models across diverse scientific fields and real-world applications, including healthcare, finance, and autonomous systems.
Papers
Normalization Perturbation: A Simple Domain Generalization Method for Real-World Domain Shifts
Qi Fan, Mattia Segu, Yu-Wing Tai, Fisher Yu, Chi-Keung Tang, Bernt Schiele, Dengxin Dai
When & How to Transfer with Transfer Learning
Adrian Tormos, Dario Garcia-Gasulla, Victor Gimenez-Abalos, Sergio Alvarez-Napagao
Self-explaining deep models with logic rule reasoning
Seungeon Lee, Xiting Wang, Sungwon Han, Xiaoyuan Yi, Xing Xie, Meeyoung Cha
Improving Out-of-Distribution Generalization by Adversarial Training with Structured Priors
Qixun Wang, Yifei Wang, Hong Zhu, Yisen Wang
Improving the Reliability for Confidence Estimation
Haoxuan Qu, Yanchao Li, Lin Geng Foo, Jason Kuen, Jiuxiang Gu, Jun Liu