Deep Learning Acceleration
Deep learning acceleration focuses on developing hardware and algorithmic techniques to significantly speed up and reduce the energy consumption of deep neural network (DNN) computations. Current research emphasizes novel architectures like processing-in-memory (PIM) systems, optical computing, and event-driven designs, alongside algorithmic optimizations such as quantization, pruning, and sparsity exploitation across various DNN models (e.g., CNNs, GNNs). These advancements are crucial for deploying DNNs on resource-constrained devices like embedded systems and edge computing platforms, enabling real-time applications in diverse fields while addressing sustainability concerns.
Papers
To Spike or Not To Spike: A Digital Hardware Perspective on Deep Learning Acceleration
Fabrizio Ottati, Chang Gao, Qinyu Chen, Giovanni Brignone, Mario R. Casu, Jason K. Eshraghian, Luciano Lavagno
A Survey on Deep Learning Hardware Accelerators for Heterogeneous HPC Platforms
Cristina Silvano, Daniele Ielmini, Fabrizio Ferrandi, Leandro Fiorin, Serena Curzel, Luca Benini, Francesco Conti, Angelo Garofalo, Cristian Zambelli, Enrico Calore, Sebastiano Fabio Schifano, Maurizio Palesi, Giuseppe Ascia, Davide Patti, Stefania Perri, Nicola Petra, Davide De Caro, Luciano Lavagno, Teodoro Urso, Valeria Cardellini, Gian Carlo Cardarilli, Robert Birke