T2V Generation
Text-to-video (T2V) generation aims to synthesize realistic videos from textual descriptions, a challenging task currently addressed using diffusion models. Research focuses on improving the compositional capabilities of these models, handling complex scenarios with multiple objects and dynamic actions, and developing robust evaluation metrics that capture video dynamics and semantic accuracy. These advancements are significant for various applications, including film production, animation, and virtual reality, and are driving improvements in both model architectures and evaluation methodologies within the broader field of generative AI.
Papers
November 13, 2024
October 8, 2024
July 2, 2024
July 1, 2024
June 10, 2024
June 6, 2024
June 5, 2024
May 9, 2024
April 25, 2024
November 24, 2023
October 30, 2023
October 11, 2023