Deep Learning Workload
Deep learning workloads encompass the computational demands of training and deploying large-scale neural networks, focusing on optimizing performance, resource utilization, and energy efficiency. Current research emphasizes efficient resource allocation and scheduling across heterogeneous hardware (CPUs, GPUs, NPUs, FPGAs), exploring techniques like model parallelism, dataflow awareness, and smart memory management to handle increasingly complex models such as Vision Transformers and large language models. These advancements are crucial for enabling the wider adoption of deep learning in various applications, from edge AI devices to massive cloud-based training clusters, by improving both speed and cost-effectiveness.
Papers
November 5, 2024
September 23, 2024
July 18, 2024
June 12, 2024
March 28, 2024
October 13, 2023
September 3, 2023
July 10, 2023
May 24, 2023
March 30, 2023
November 19, 2022
November 10, 2022
September 13, 2022
July 23, 2022
June 3, 2022
May 26, 2022
February 16, 2022
January 27, 2022
December 2, 2021