Pre Trained Vision Transformer
Pre-trained Vision Transformers (ViTs) leverage the power of large-scale pre-training to efficiently adapt to diverse downstream computer vision tasks, focusing on parameter and computational efficiency. Current research emphasizes methods like low-rank adaptation, visual prompt tuning, and multiple-exit strategies to minimize the number of trainable parameters and optimize inference speed, particularly for resource-constrained environments like onboard satellite processing and mobile devices. These advancements are significantly impacting various fields, including remote sensing, medical image analysis, and robotics, by enabling high-accuracy visual recognition with reduced computational demands.
Papers
November 11, 2024
November 8, 2024
October 30, 2024
October 27, 2024
September 23, 2024
September 21, 2024
September 5, 2024
August 13, 2024
August 3, 2024
July 16, 2024
July 9, 2024
July 5, 2024
June 24, 2024
June 16, 2024
June 15, 2024
June 5, 2024
May 31, 2024
May 28, 2024
May 24, 2024
May 23, 2024