Train Test
Train-test discrepancies, where differences between training and testing data negatively impact model performance, are a significant challenge across various machine learning domains. Current research focuses on mitigating these discrepancies through techniques like developing scale-invariant models, refining data preprocessing to reduce class overlap, and improving the generalization capabilities of models, particularly in long-sequence processing (e.g., LLMs) and weakly supervised learning. Addressing these issues is crucial for improving the reliability and robustness of machine learning models in real-world applications, ranging from information retrieval and image recognition to natural language processing.
Papers
October 10, 2024
October 2, 2024
April 1, 2024
February 29, 2024
October 14, 2023