DNN Workload
DNN workload research focuses on optimizing the execution of deep neural networks (DNNs), encompassing single and multi-DNN scenarios, across diverse hardware platforms ranging from mobile devices to large data centers. Current efforts concentrate on improving efficiency through techniques like model sparsification, novel scheduling algorithms (e.g., dynamic and static scheduling), and hardware acceleration using specialized architectures (e.g., spiking neural networks, optical computing, and tensor cores). These advancements aim to address critical bottlenecks in performance, energy consumption, and latency, ultimately impacting the scalability and real-world applicability of AI systems in various domains.
Papers
October 15, 2024
September 25, 2024
September 2, 2024
July 30, 2024
July 19, 2024
May 5, 2024
December 16, 2023
October 17, 2023
July 23, 2023
July 16, 2023
July 6, 2023
June 21, 2022
May 19, 2022
May 17, 2022
December 18, 2021