Data Normalisation
Data normalization is a crucial preprocessing step in machine learning, aiming to standardize data features for improved model performance and fairness. Current research emphasizes the impact of normalization choices on model explainability, particularly in sensitive domains like medicine, and explores optimal normalization strategies for various data types, including text and biomedical data, often comparing different distance functions and their effectiveness in different contexts. These efforts are significant because appropriate normalization techniques are essential for reliable model training, accurate evaluation (e.g., avoiding biases in metrics like nDCG), and ensuring fair and generalizable results across diverse datasets.
Papers
November 18, 2024
August 30, 2024
February 19, 2024
July 27, 2023
June 30, 2023
March 9, 2023
October 3, 2022
June 13, 2022
June 7, 2022
March 29, 2022
March 7, 2022
March 1, 2022
February 2, 2022