Training Data Influence
Training data influence research investigates how individual training examples impact the performance and behavior of machine learning models, particularly focusing on identifying and mitigating the effects of noisy, biased, or otherwise problematic data. Current research employs various techniques, including influence functions and gradient-based methods, across diverse model architectures such as large language models (LLMs), diffusion models, and deep learning models for image classification. Understanding and addressing training data influence is crucial for improving model robustness, fairness, and explainability, leading to more reliable and trustworthy AI systems across numerous applications.
Papers
October 24, 2024
October 18, 2024
September 5, 2024
April 11, 2024
March 15, 2024
January 26, 2024
October 7, 2023
August 17, 2023
June 3, 2023
December 9, 2022
July 28, 2022
February 24, 2022