Pre Trained Vision Transformer
Pre-trained Vision Transformers (ViTs) leverage the power of large-scale pre-training to efficiently adapt to diverse downstream computer vision tasks, focusing on parameter and computational efficiency. Current research emphasizes methods like low-rank adaptation, visual prompt tuning, and multiple-exit strategies to minimize the number of trainable parameters and optimize inference speed, particularly for resource-constrained environments like onboard satellite processing and mobile devices. These advancements are significantly impacting various fields, including remote sensing, medical image analysis, and robotics, by enabling high-accuracy visual recognition with reduced computational demands.
Papers
March 1, 2023
February 28, 2023
February 8, 2023
February 6, 2023
February 1, 2023
January 24, 2023
December 6, 2022
November 23, 2022
November 1, 2022
October 23, 2022
October 13, 2022
October 6, 2022
July 27, 2022
July 10, 2022
June 22, 2022
May 30, 2022
May 26, 2022
May 20, 2022
April 8, 2022
March 10, 2022