Train Test
Train-test discrepancies, where differences between training and testing data negatively impact model performance, are a significant challenge across various machine learning domains. Current research focuses on mitigating these discrepancies through techniques like developing scale-invariant models, refining data preprocessing to reduce class overlap, and improving the generalization capabilities of models, particularly in long-sequence processing (e.g., LLMs) and weakly supervised learning. Addressing these issues is crucial for improving the reliability and robustness of machine learning models in real-world applications, ranging from information retrieval and image recognition to natural language processing.
Papers
November 20, 2024
October 10, 2024
October 2, 2024
April 1, 2024
February 29, 2024
October 14, 2023