Perception Aware
Perception-aware systems aim to build artificial intelligence that understands and interacts with the world in ways similar to humans, focusing on robust and accurate perception across various modalities (vision, language, etc.). Current research emphasizes improving the accuracy and efficiency of perception models, particularly through advancements in vision-language models (VLMs) and the development of novel algorithms for dynamic resolution processing, multimodal data fusion, and uncertainty quantification. This field is crucial for advancing robotics, autonomous driving, and human-computer interaction, with applications ranging from improved object recognition and scene understanding to more natural and intuitive interfaces.
Papers
FlowAct: A Proactive Multimodal Human-robot Interaction System with Continuous Flow of Perception and Modular Action Sub-systems
Timothée Dhaussy, Bassam Jabaian, Fabrice Lefèvre
On the Benefits of Visual Stabilization for Frame- and Event-based Perception
Juan Pablo Rodriguez-Gomez, Jose Ramiro Martinez-de Dios, Anibal Ollero, Guillermo Gallego