Adverse Weather

Adverse weather significantly degrades the performance of computer vision systems used in autonomous driving and other applications. Current research focuses on developing robust models, often employing transformer networks, diffusion models, and contrastive learning, to improve image and LiDAR point cloud processing in challenging conditions like rain, fog, snow, and low light. This work emphasizes improving the accuracy and safety of perception systems by mitigating the effects of adverse weather on object detection, semantic segmentation, and other crucial tasks. The resulting advancements have significant implications for the safety and reliability of autonomous vehicles and other applications relying on robust environmental perception.

Papers