Supervised ImageNet
Supervised ImageNet research focuses on improving image classification models by leveraging the massive ImageNet dataset. Current efforts concentrate on enhancing data curation strategies, developing more efficient training methods (including exploring alternative architectures like binary neural networks and leveraging self-supervised learning), and addressing challenges like dataset bias and the need for explainable AI. These advancements are crucial for improving the accuracy, efficiency, and trustworthiness of computer vision systems across various applications, from medical imaging to agricultural technology.
Papers
CPPE-5: Medical Personal Protective Equipment Dataset
Rishit Dagli, Ali Mustufa Shaikh
An Experimental Study of the Impact of Pre-training on the Pruning of a Convolutional Neural Network
Nathan Hubens, Matei Mancas, Bernard Gosselin, Marius Preda, Titus Zaharia
A Comparative Analysis of Machine Learning Approaches for Automated Face Mask Detection During COVID-19
Junaed Younus Khan, Md Abdullah Al Alamin
N-ImageNet: Towards Robust, Fine-Grained Object Recognition with Event Cameras
Junho Kim, Jaehyeok Bae, Gangin Park, Dongsu Zhang, Young Min Kim
Temporally Resolution Decrement: Utilizing the Shape Consistency for Higher Computational Efficiency
Tianshu Xie, Xuan Cheng, Minghui Liu, Jiali Deng, Xiaomin Wang, Ming Liu
Improving Differentiable Architecture Search with a Generative Model
Ruisi Zhang, Youwei Liang, Sai Ashish Somayajula, Pengtao Xie
Task2Sim : Towards Effective Pre-training and Transfer from Synthetic Data
Samarth Mishra, Rameswar Panda, Cheng Perng Phoo, Chun-Fu Chen, Leonid Karlinsky, Kate Saenko, Venkatesh Saligrama, Rogerio S. Feris
AdaViT: Adaptive Vision Transformers for Efficient Image Recognition
Lingchen Meng, Hengduo Li, Bor-Chun Chen, Shiyi Lan, Zuxuan Wu, Yu-Gang Jiang, Ser-Nam Lim
Deep Models for Visual Sentiment Analysis of Disaster-related Multimedia Content
Khubaib Ahmad, Muhammad Asif Ayub, Kashif Ahmad, Ala Al-Fuqaha, Nasir Ahmad
A Unified Pruning Framework for Vision Transformers
Hao Yu, Jianxin Wu
Pyramid Adversarial Training Improves ViT Performance
Charles Herrmann, Kyle Sargent, Lu Jiang, Ramin Zabih, Huiwen Chang, Ce Liu, Dilip Krishnan, Deqing Sun
DAFormer: Improving Network Architectures and Training Strategies for Domain-Adaptive Semantic Segmentation
Lukas Hoyer, Dengxin Dai, Luc Van Gool
Searching the Search Space of Vision Transformer
Minghao Chen, Kan Wu, Bolin Ni, Houwen Peng, Bei Liu, Jianlong Fu, Hongyang Chao, Haibin Ling
Enhanced Transfer Learning Through Medical Imaging and Patient Demographic Data Fusion
Spencer A. Thomas