Long Term Video
Long-term video understanding aims to develop computational models capable of analyzing and interpreting video sequences extending beyond short clips, focusing on comprehending complex, temporally extended events and actions. Current research emphasizes the development of robust multimodal models, often incorporating transformer architectures and diffusion models, to handle long-range dependencies and noisy data inherent in extended video content. This field is crucial for advancing video question answering, object segmentation, and other video-related tasks, ultimately impacting applications ranging from video search and summarization to robotics and virtual reality.
Papers
November 20, 2024
November 19, 2024
October 21, 2024
September 12, 2024
July 22, 2024
June 3, 2024
April 30, 2024
April 4, 2024
March 26, 2024
January 30, 2024
December 8, 2023
October 16, 2023
August 20, 2023
July 18, 2023
March 31, 2023
March 30, 2023
November 18, 2022
November 17, 2022
October 5, 2022