Supervised ImageNet
Supervised ImageNet research focuses on improving image classification models by leveraging the massive ImageNet dataset. Current efforts concentrate on enhancing data curation strategies, developing more efficient training methods (including exploring alternative architectures like binary neural networks and leveraging self-supervised learning), and addressing challenges like dataset bias and the need for explainable AI. These advancements are crucial for improving the accuracy, efficiency, and trustworthiness of computer vision systems across various applications, from medical imaging to agricultural technology.
Papers
Breaking the Architecture Barrier: A Method for Efficient Knowledge Transfer Across Networks
Maciej A. Czyzewski, Daniel Nowak, Kamil Piechowiak
Swin MAE: Masked Autoencoders for Small Datasets
Zi'an Xu, Yin Dai, Fayu Liu, Weibing Chen, Yue Liu, Lifu Shi, Sheng Liu, Yuhang Zhou
OVO: One-shot Vision Transformer Search with Online distillation
Zimian Wei, Hengyue Pan, Xin Niu, Dongsheng Li
CLIP Itself is a Strong Fine-tuner: Achieving 85.7% and 88.0% Top-1 Accuracy with ViT-B and ViT-L on ImageNet
Xiaoyi Dong, Jianmin Bao, Ting Zhang, Dongdong Chen, Shuyang Gu, Weiming Zhang, Lu Yuan, Dong Chen, Fang Wen, Nenghai Yu
Robust Perception through Equivariance
Chengzhi Mao, Lingyu Zhang, Abhishek Joshi, Junfeng Yang, Hao Wang, Carl Vondrick
Carpet-bombing patch: attacking a deep network without usual requirements
Pol Labarbarie, Adrien Chan-Hon-Tong, Stéphane Herbin, Milad Leyli-Abadi