Direct Convolution
Direct convolution, a fundamental operation in convolutional neural networks (CNNs), is being actively researched to improve efficiency and accuracy across diverse applications. Current efforts focus on optimizing convolution through architectural innovations like dilated convolutions, attention mechanisms integrated with convolutions, and novel data layouts for improved hardware performance, as well as exploring alternatives to traditional convolutions, such as using semirings or table lookups. These advancements aim to enhance the speed and accuracy of CNNs for tasks ranging from medical image analysis and object detection to speech processing and large language model efficiency, ultimately impacting various scientific fields and practical applications.
Papers
QuadConv: Quadrature-Based Convolutions with Applications to Non-Uniform PDE Data Compression
Kevin Doherty, Cooper Simpson, Stephen Becker, Alireza Doostan
LiCo-Net: Linearized Convolution Network for Hardware-efficient Keyword Spotting
Haichuan Yang, Zhaojun Yang, Li Wan, Biqiao Zhang, Yangyang Shi, Yiteng Huang, Ivaylo Enchev, Limin Tang, Raziel Alvarez, Ming Sun, Xin Lei, Raghuraman Krishnamoorthi, Vikas Chandra
Deformably-Scaled Transposed Convolution
Stefano B. Blumberg, Daniele Raví, Mou-Cheng Xu, Matteo Figini, Iasonas Kokkinos, Daniel C. Alexander
Packed-Ensembles for Efficient Uncertainty Estimation
Olivier Laurent, Adrien Lafage, Enzo Tartaglione, Geoffrey Daniel, Jean-Marc Martinez, Andrei Bursuc, Gianni Franchi