Large Scale Benchmark
Large-scale benchmarks are datasets designed to rigorously evaluate the performance of machine learning models across diverse and challenging tasks, pushing the boundaries of model capabilities. Current research focuses on developing benchmarks for various domains, including fluid dynamics, log parsing, image manipulation detection, and various aspects of video and image analysis, often employing deep learning architectures like transformers and convolutional neural networks. These benchmarks are crucial for advancing the field by providing standardized evaluation metrics and facilitating the development of more robust and generalizable models with significant implications for diverse applications ranging from medical imaging to autonomous systems.
Papers
UrbanBIS: a Large-scale Benchmark for Fine-grained Urban Building Instance Segmentation
Guoqing Yang, Fuyou Xue, Qi Zhang, Ke Xie, Chi-Wing Fu, Hui Huang
ANetQA: A Large-scale Benchmark for Fine-grained Compositional Reasoning over Untrimmed Videos
Zhou Yu, Lixiang Zheng, Zhou Zhao, Fei Wu, Jianping Fan, Kui Ren, Jun Yu