Knowledge Comprehension Capability
Knowledge comprehension capability research focuses on developing and evaluating systems, primarily large language models (LLMs), that can accurately understand and reason with textual and multimodal information. Current research emphasizes improving LLMs' ability to handle complex contexts, diverse linguistic styles, and nuanced information, often employing techniques like multimodal learning, progressive comprehension networks, and coupled comprehension-generation architectures. This field is crucial for advancing AI's ability to interact meaningfully with humans and process information from diverse sources, with implications for applications ranging from education and healthcare to information retrieval and knowledge graph construction.
Papers
Long-Span Question-Answering: Automatic Question Generation and QA-System Ranking via Side-by-Side Evaluation
Bernd Bohnet, Kevin Swersky, Rosanne Liu, Pranjal Awasthi, Azade Nova, Javier Snaider, Hanie Sedghi, Aaron T Parisi, Michael Collins, Angeliki Lazaridou, Orhan Firat, Noah Fiedel
StrucTexTv3: An Efficient Vision-Language Model for Text-rich Image Perception, Comprehension, and Beyond
Pengyuan Lyu, Yulin Li, Hao Zhou, Weihong Ma, Xingyu Wan, Qunyi Xie, Liang Wu, Chengquan Zhang, Kun Yao, Errui Ding, Jingdong Wang