Direct Convolution
Direct convolution, a fundamental operation in convolutional neural networks (CNNs), is being actively researched to improve efficiency and accuracy across diverse applications. Current efforts focus on optimizing convolution through architectural innovations like dilated convolutions, attention mechanisms integrated with convolutions, and novel data layouts for improved hardware performance, as well as exploring alternatives to traditional convolutions, such as using semirings or table lookups. These advancements aim to enhance the speed and accuracy of CNNs for tasks ranging from medical image analysis and object detection to speech processing and large language model efficiency, ultimately impacting various scientific fields and practical applications.
Papers
RISurConv: Rotation Invariant Surface Attention-Augmented Convolutions for 3D Point Cloud Classification and Segmentation
Zhiyuan Zhang, Licheng Yang, Zhiyu Xiang
ControlNeXt: Powerful and Efficient Control for Image and Video Generation
Bohao Peng, Jian Wang, Yuechen Zhang, Wenbo Li, Ming-Chang Yang, Jiaya Jia
Layer-Specific Optimization: Sensitivity Based Convolution Layers Basis Search
Vasiliy Alekseev, Ilya Lukashevich, Ilia Zharikov, Ilya Vasiliev
TF-Locoformer: Transformer with Local Modeling by Convolution for Speech Separation and Enhancement
Kohei Saijo, Gordon Wichern, François G. Germain, Zexu Pan, Jonathan Le Roux
Don't Think It Twice: Exploit Shift Invariance for Efficient Online Streaming Inference of CNNs
Christodoulos Kechris, Jonathan Dan, Jose Miranda, David Atienza