Text Only Language Model
Text-only language models (LMs) are being extensively researched to expand their capabilities beyond text processing, focusing on integrating visual information and improving robustness. Current efforts involve adapting existing LM architectures to handle multimodal inputs (e.g., images and text) through techniques like linear projections between image and text spaces, or by leveraging the LMs' strong language understanding to guide multimodal generation. This research is significant because it allows leveraging the power of pre-trained LMs for tasks requiring both linguistic and visual understanding, potentially leading to advancements in areas like machine translation, visual question answering, and interactive chatbot development.
Papers
September 26, 2024
May 21, 2024
March 20, 2024
October 15, 2023
September 5, 2023
January 31, 2023
December 20, 2022
October 27, 2022