Deep Model
Deep models, encompassing a broad range of neural network architectures, aim to learn complex patterns from data for various tasks like image classification, time series forecasting, and system identification. Current research emphasizes improving efficiency (e.g., through constant-time learning algorithms and layer caching), enhancing explainability (e.g., via gradient-free methods), and mitigating issues like bias and memorization. These advancements are significant because they improve the reliability, trustworthiness, and applicability of deep models across diverse scientific fields and real-world applications, including healthcare, finance, and autonomous systems.
Papers
Efficient model compression with Random Operation Access Specific Tile (ROAST) hashing
Aditya Desai, Keren Zhou, Anshumali Shrivastava
UFO: Unified Feature Optimization
Teng Xi, Yifan Sun, Deli Yu, Bi Li, Nan Peng, Gang Zhang, Xinyu Zhang, Zhigang Wang, Jinwen Chen, Jian Wang, Lufei Liu, Haocheng Feng, Junyu Han, Jingtuo Liu, Errui Ding, Jingdong Wang
ACT-Net: Asymmetric Co-Teacher Network for Semi-supervised Memory-efficient Medical Image Segmentation
Ziyuan Zhao, Andong Zhu, Zeng Zeng, Bharadwaj Veeravalli, Cuntai Guan
PoF: Post-Training of Feature Extractor for Improving Generalization
Ikuro Sato, Ryota Yamada, Masayuki Tanaka, Nakamasa Inoue, Rei Kawakami