Deep Neural Network
Deep neural networks (DNNs) are complex computational models aiming to mimic the human brain's learning capabilities, primarily focusing on achieving high accuracy and efficiency in various tasks. Current research emphasizes understanding DNN training dynamics, including phenomena like neural collapse and the impact of architectural choices (e.g., convolutional, transformer, and operator networks) and training strategies (e.g., weight decay, knowledge distillation, active learning). This understanding is crucial for improving DNN performance, robustness (including against adversarial attacks and noisy data), and resource efficiency in diverse applications ranging from image recognition and natural language processing to scientific modeling and edge computing.
Papers
Subword Embedding from Bytes Gains Privacy without Sacrificing Accuracy and Complexity
Mengjiao Zhang, Jia Xu
Addressing Spectral Bias of Deep Neural Networks by Multi-Grade Deep Learning
Ronglong Fang, Yuesheng Xu
DeepVigor+: Scalable and Accurate Semi-Analytical Fault Resilience Analysis for Deep Neural Network
Mohammad Hasan Ahmadilivani, Jaan Raik, Masoud Daneshtalab, Maksim Jenihhin
AutoAL: Automated Active Learning with Differentiable Query Strategy Search
Yifeng Wang, Xueying Zhan, Siyu Huang
OAH-Net: A Deep Neural Network for Hologram Reconstruction of Off-axis Digital Holographic Microscope
Wei Liu, Kerem Delikoyun, Qianyu Chen, Alperen Yildiz, Si Ko Myo, Win Sen Kuan, John Tshon Yit Soong, Matthew Edward Cove, Oliver Hayden, Hweekuan Lee
Shavette: Low Power Neural Network Acceleration via Algorithm-level Error Detection and Undervolting
Mikael Rinkinen, Lauri Koskinen, Olli Silven, Mehdi Safarpour
FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Zhenheng Tang, Xueze Kang, Yiming Yin, Xinglin Pan, Yuxin Wang, Xin He, Qiang Wang, Rongfei Zeng, Kaiyong Zhao, Shaohuai Shi, Amelie Chi Zhou, Bo Li, Bingsheng He, Xiaowen Chu
The Bayesian Confidence (BACON) Estimator for Deep Neural Networks
Patrick D. Kee, Max J. Brown, Jonathan C. Rice, Christian A. Howell
Feature Clipping for Uncertainty Calibration
Linwei Tao, Minjing Dong, Chang Xu
DiffGAN: A Test Generation Approach for Differential Testing of Deep Neural Networks
Zohreh Aghababaeyan, Manel Abdellatif, Lionel Briand, Ramesh S
Leveraging Multi-Temporal Sentinel 1 and 2 Satellite Data for Leaf Area Index Estimation With Deep Learning
Clement Wang, Antoine Debouchage, Valentin Goldité, Aurélien Wery, Jules Salzinger
Advancing the Understanding of Fixed Point Iterations in Deep Neural Networks: A Detailed Analytical Study
Yekun Ke, Xiaoyu Li, Yingyu Liang, Zhenmei Shi, Zhao Song
Towards Understanding Why FixMatch Generalizes Better Than Supervised Learning
Jingyang Li, Jiachun Pan, Vincent Y. F. Tan, Kim-Chuan Toh, Pan Zhou
Statistical Properties of Deep Neural Networks with Dependent Data
Chad Brown
Non-convergence to global minimizers in data driven supervised deep learning: Adam and stochastic gradient descent optimization provably fail to converge to global minimizers in the training of deep neural networks with ReLU activation
Sonja Hannibal, Arnulf Jentzen, Do Minh Thang