Resolution Transformer
Resolution Transformers leverage the power of transformer architectures to process data at multiple resolutions, aiming to improve efficiency and accuracy in various computer vision and time-series tasks. Current research focuses on developing efficient multi-resolution architectures, such as those employing dual knowledge distillation or cross-resolution attention mechanisms, to balance computational cost with performance gains. This approach is proving impactful across diverse applications, including object tracking, robotic control, point cloud completion, and semantic segmentation, by enabling the integration of both local and global contextual information.
Papers
May 27, 2024
January 25, 2024
January 3, 2024
November 7, 2023
June 12, 2023
May 24, 2023
March 13, 2023
January 8, 2023
December 15, 2022