Data Parallelism
Data parallelism accelerates computation by distributing data across multiple processors, enabling faster training of large models in machine learning and scientific computing. Current research focuses on optimizing data partitioning strategies for various architectures (e.g., K-means clustering, deep neural networks, large language models), including hybrid approaches that combine synchronous and asynchronous methods to improve efficiency and scalability. These advancements are crucial for tackling increasingly complex problems in fields like AI and scientific simulation, where the size and complexity of datasets demand efficient parallel processing techniques.
Papers
June 27, 2024
March 27, 2024
March 13, 2024
February 28, 2024
December 8, 2023
May 22, 2023
February 17, 2023
November 19, 2022
November 11, 2022
December 21, 2021
December 6, 2021