Image Captioning Model
Image captioning models automatically generate textual descriptions of images, aiming to create captions that are both accurate and engaging. Current research focuses on improving caption quality through techniques like direct optimization using CLIP scores, developing more efficient architectures (e.g., those based on Fourier transforms), and enhancing robustness against adversarial attacks. These advancements are significant for various applications, including accessibility tools, content creation, and improving the performance of larger vision-language models, while also raising important considerations around AI safety and ethical deployment.
Papers
September 27, 2023
September 26, 2023
September 5, 2023
July 31, 2023
May 4, 2023
November 4, 2022
October 19, 2022
September 17, 2022
July 20, 2022
May 26, 2022
April 8, 2022
December 9, 2021