Bilingual Multimodal
Bilingual multimodal research focuses on developing large language models (LLMs) capable of understanding and processing both textual and visual information in multiple languages. Current efforts concentrate on creating and utilizing large bilingual datasets to train models, employing techniques like contrastive learning and incorporating visual receptors to enhance image-text alignment. These advancements are improving performance on complex tasks like scientific problem-solving and chemical reasoning, demonstrating the potential for broader applications in scientific research and beyond.
Papers
October 17, 2024
August 14, 2024
March 18, 2024
February 21, 2024
August 24, 2023