CUDA Implementation
CUDA implementation focuses on leveraging the parallel processing power of GPUs to accelerate computationally intensive tasks, primarily in scientific computing and machine learning. Current research emphasizes optimizing CUDA kernels for specific algorithms (e.g., sparse matrix multiplication, convolutional neural networks) and improving the reproducibility and efficiency of CUDA-based applications, including addressing randomness and memory management challenges. These advancements significantly impact various fields by enabling faster training of machine learning models, real-time simulations, and high-resolution rendering, ultimately accelerating scientific discovery and technological innovation.
Papers
October 27, 2024
September 19, 2024
September 17, 2024
September 11, 2024
August 19, 2024
August 15, 2024
June 30, 2024
June 25, 2024
June 4, 2024
March 26, 2024
February 21, 2024
February 12, 2024
November 1, 2023
September 12, 2023
August 17, 2023
July 24, 2023
June 27, 2023
June 6, 2023
May 15, 2023