Event Based Action Recognition
Event-based action recognition focuses on using data from event cameras—sensors that record changes in pixel intensity rather than full frames—to identify human actions. Current research emphasizes developing robust and efficient algorithms, including convolutional neural networks, spiking neural networks, and transformer-based architectures, often coupled with novel data representations like point clouds and hypergraphs, to improve accuracy and reduce power consumption. This field is significant due to the potential for low-power, high-speed action recognition in applications like augmented reality and human-computer interaction, driven by the creation of larger, more comprehensive datasets for model training and evaluation.
Papers
Helios: An extremely low power event-based gesture recognition for always-on smart eyewear
Prarthana Bhattacharyya, Joshua Mitton, Ryan Page, Owen Morgan, Ben Menzies, Gabriel Homewood, Kemi Jacobs, Paolo Baesso, David Trickett, Chris Mair, Taru Muhonen, Rory Clark, Louis Berridge, Richard Vigars, Iain Wallace
DailyDVS-200: A Comprehensive Benchmark Dataset for Event-Based Action Recognition
Qi Wang, Zhou Xu, Yuming Lin, Jingtao Ye, Hongsheng Li, Guangming Zhu, Syed Afaq Ali Shah, Mohammed Bennamoun, Liang Zhang