Benchmark Datasets
Benchmark datasets are crucial for evaluating the performance of machine learning models across diverse tasks, from natural language processing to image analysis and graph classification. Current research emphasizes the need for more robust and representative datasets, addressing issues like data leakage, bias, and distribution mismatches that can skew results and hinder fair comparisons between models. This focus on improved dataset quality is vital for ensuring the reliability of model evaluations and driving progress in the development of more accurate and generalizable algorithms, ultimately impacting the trustworthiness and practical applicability of AI systems.
Papers
A Survey of Small Language Models
Chien Van Nguyen, Xuan Shen, Ryan Aponte, Yu Xia, Samyadeep Basu, Zhengmian Hu, Jian Chen, Mihir Parmar, Sasidhar Kunapuli, Joe Barrow, Junda Wu, Ashish Singh, Yu Wang, Jiuxiang Gu, Franck Dernoncourt, Nesreen K. Ahmed, Nedim Lipka, Ruiyi Zhang, Xiang Chen, Tong Yu, Sungchul Kim, Hanieh Deilamsalehy, Namyong Park, Mike Rimer, Zhehao Zhang, Huanrui Yang, Ryan A. Rossi, Thien Huu Nguyen
On Occlusions in Video Action Detection: Benchmark Datasets And Training Recipes
Rajat Modi, Vibhav Vineet, Yogesh Singh Rawat