Editing Capability
Image and video editing using generative models is a rapidly advancing field focused on improving controllability, efficiency, and fidelity. Current research emphasizes developing novel algorithms, such as score distillation sampling and various attention mechanisms, to enhance the precision and speed of editing while preserving original content integrity. These advancements leverage diffusion models, GANs, and increasingly, large language models for text-guided editing, aiming to create more intuitive and powerful tools for content creation. The resulting improvements have significant implications for various applications, including creative design, media production, and potentially even scientific visualization.
Papers
SVGDreamer++: Advancing Editability and Diversity in Text-Guided SVG Generation
Ximing Xing, Qian Yu, Chuang Wang, Haitao Zhou, Jing Zhang, Dong Xu
DreamMix: Decoupling Object Attributes for Enhanced Editability in Customized Image Inpainting
Yicheng Yang, Pengxiang Li, Lu Zhang, Liqian Ma, Ping Hu, Siyu Du, Yunzhi Zhuge, Xu Jia, Huchuan Lu