FPGA Accelerator
FPGA accelerators are hardware devices designed to significantly speed up the execution of computationally intensive algorithms, particularly in machine learning. Current research focuses on optimizing these accelerators for various neural network architectures, including transformers (like those used in large language models), convolutional neural networks, and spiking neural networks, often employing techniques like pipelining, systolic arrays, and mixed-scheme quantization to maximize throughput and energy efficiency. This work is crucial for deploying advanced AI models on resource-constrained devices like edge computers and embedded systems, enabling faster and more power-efficient applications in diverse fields such as computer vision, natural language processing, and robotics.
Papers
FAMOUS: Flexible Accelerator for the Attention Mechanism of Transformer on UltraScale+ FPGAs
Ehsan Kabir, Md. Arafat Kabir, Austin R.J. Downey, Jason D. Bakos, David Andrews, Miaoqing Huang
ProTEA: Programmable Transformer Encoder Acceleration on FPGA
Ehsan Kabir, Jason D. Bakos, David Andrews, Miaoqing Huang