Memory Efficient
Memory-efficient deep learning focuses on developing and optimizing neural network models and training algorithms to minimize memory consumption, enabling deployment on resource-constrained devices and scaling to larger models. Current research emphasizes techniques like neural architecture search (NAS) to design inherently efficient architectures (e.g., optimized convolutional and transformer networks, memory-efficient Graph Neural Networks), low-rank approximations, mixed-precision training, and novel optimization algorithms (e.g., variants of Adam, Shampoo) that reduce memory overhead during both training and inference. These advancements are crucial for expanding the accessibility and applicability of deep learning across diverse hardware platforms and datasets, particularly in edge computing and large language model training.
Papers
An FPGA-Based Reconfigurable Accelerator for Convolution-Transformer Hybrid EfficientViT
Haikuo Shao, Huihong Shi, Wendong Mao, Zhongfeng Wang
Mixed-precision Supernet Training from Vision Foundation Models using Low Rank Adapter
Yuiko Sakuma, Masakazu Yoshimura, Junji Otsuka, Atsushi Irie, Takeshi Ohashi