Background Generation
Background generation in computer vision focuses on creating realistic and contextually appropriate backgrounds for images and videos, often to enhance image editing, object detection, or video synthesis. Current research emphasizes improving control over background generation, including precise spatial relationships between foreground and background elements, and mitigating biases in background representation that can lead to inaccurate or stereotypical outputs. This is achieved through various techniques, such as leveraging diffusion models, transformers, and contrastive learning, often in conjunction with attention mechanisms and novel data augmentation strategies. Advances in this area have significant implications for numerous applications, including image editing, e-commerce, and the development of more robust and generalizable computer vision models.
Papers
Salient Object-Aware Background Generation using Text-Guided Diffusion Models
Amir Erfan Eshratifar, Joao V. B. Soares, Kapil Thadani, Shaunak Mishra, Mikhail Kuznetsov, Yueh-Ning Ku, Paloma de Juan
ViFu: Multiple 360$^\circ$ Objects Reconstruction with Clean Background via Visible Part Fusion
Tianhan Xu, Takuya Ikeda, Koichi Nishiwaki