Equivariant Network
Equivariant networks are neural networks designed to leverage data symmetries, improving efficiency and generalization by ensuring consistent behavior under specific transformations (e.g., rotations, translations). Current research focuses on developing efficient and expressive architectures, such as those based on polynomial formulations, Kolmogorov-Arnold networks, and various group convolutional approaches, addressing challenges in optimization and achieving equivariance for diverse groups and data types (e.g., point clouds, graphs, images). This field is significant because it allows for the development of more data-efficient and robust models across numerous applications, including 3D object reconstruction, particle physics, and molecular dynamics simulations.
Papers
Approximate Equivariance in Reinforcement Learning
Jung Yeon Park, Sujay Bhatt, Sihan Zeng, Lawson L.S. Wong, Alec Koppel, Sumitra Ganesh, Robin Walters
Equivariant Graph Network Approximations of High-Degree Polynomials for Force Field Prediction
Zhao Xu, Haiyang Yu, Montgomery Bohde, Shuiwang Ji
Harmformer: Harmonic Networks Meet Transformers for Continuous Roto-Translation Equivariance
Tomáš Karella, Adam Harmanec, Jan Kotera, Jan Blažek, Filip Šroubek