Distilled Dataset
Dataset distillation aims to create significantly smaller, synthetic datasets that retain the essential information of much larger original datasets, enabling faster and more efficient training of machine learning models. Current research focuses on improving the quality and robustness of these distilled datasets, exploring techniques like matching-based methods, diffusion models, and the strategic use of soft labels to address issues such as class imbalance and cross-architecture generalization. This field is significant because it offers solutions to the computational and storage challenges posed by massive datasets, impacting areas like federated learning, resource-constrained applications, and model compression.
Papers
October 20, 2024
September 26, 2024
August 24, 2024
August 22, 2024
August 6, 2024
July 21, 2024
July 15, 2024
June 21, 2024
June 15, 2024
June 6, 2024
May 23, 2024
May 22, 2024
April 26, 2024
April 11, 2024
March 20, 2024
March 18, 2024
March 15, 2024
March 6, 2024
February 20, 2024