Biased Feature
Biased features in machine learning models refer to spurious correlations in training data that lead models to rely on irrelevant attributes (e.g., race, gender) for predictions, rather than task-relevant information. Current research focuses on identifying and mitigating these biases through various techniques, including adversarial training, feature orthogonalization, and dataset refinement methods applied to diverse model architectures like CNNs and LLMs. Addressing biased features is crucial for ensuring fairness, improving model generalization, and building trustworthy AI systems across various applications, from image recognition to natural language processing.
Papers
November 1, 2024
October 25, 2024
October 19, 2024
August 18, 2024
August 10, 2024
June 18, 2024
May 23, 2024
March 21, 2024
February 21, 2024
November 2, 2023
November 1, 2023
August 13, 2023
July 20, 2023
May 11, 2023
May 6, 2023
December 11, 2022
November 23, 2022
November 4, 2022
November 2, 2022