Cross Modality
Cross-modality research focuses on integrating information from different data types (e.g., images, text, audio) to improve model performance and understanding. Current research emphasizes developing robust methods for handling inconsistencies between modalities, particularly using techniques like contrastive learning, generative adversarial networks (GANs), and vision transformers, often within frameworks of unsupervised domain adaptation or self-training. This field is significant for advancing medical image analysis (e.g., improved segmentation and diagnosis), autonomous driving, and other applications requiring the fusion of heterogeneous data sources, ultimately leading to more accurate and reliable systems.
Papers
Language Guided Domain Generalized Medical Image Segmentation
Shahina Kunhimon, Muzammal Naseer, Salman Khan, Fahad Shahbaz Khan
Diffusion based Zero-shot Medical Image-to-Image Translation for Cross Modality Segmentation
Zihao Wang, Yingyu Yang, Yuzhou Chen, Tingting Yuan, Maxime Sermesant, Herve Delingette, Ona Wu