Tensor Processing Unit
Tensor Processing Units (TPUs) are specialized hardware accelerators designed to significantly speed up machine learning computations, primarily focusing on the matrix multiplications central to deep neural networks. Current research emphasizes optimizing TPU performance for various model architectures, including convolutional neural networks (CNNs), graph neural networks (GNNs), and generative adversarial networks (GANs), through techniques like precision reduction, reconfigurable dataflows, and algorithm-hardware co-design. This focus on efficiency and scalability makes TPUs crucial for accelerating diverse applications, from real-time object detection in robotics to large-scale graph embeddings and high-throughput AI training.
Papers
October 5, 2024
September 17, 2024
July 11, 2024
April 29, 2024
November 7, 2023
September 29, 2023
September 16, 2023
July 26, 2023
May 4, 2023
April 18, 2023
November 21, 2022
October 23, 2022
October 21, 2022
June 28, 2022
April 9, 2022
April 7, 2022
January 1, 2022
November 11, 2021
November 8, 2021