Static Deep Neural Network
Static deep neural networks (DNNs), while powerful, face limitations in adapting to dynamic data or resource constraints. Current research focuses on enhancing their efficiency and robustness through techniques like compressed replay for continual learning, optimized compilation frameworks for dynamic DNN variants, and novel architectures such as hypernetworks for improved performance at lower computational costs. These advancements aim to improve the practicality and reliability of DNNs across diverse applications, particularly in resource-limited environments and scenarios requiring adaptation to evolving data streams.
Papers
July 17, 2024
May 29, 2024
February 29, 2024
February 23, 2024
January 17, 2024
August 17, 2023
May 27, 2023
October 17, 2022