Equivariant non-linear maps for neural networks on homogeneous spaces
Preprint, 2025

This paper presents a novel framework for non-linear equivariant neural network layers on homogeneous spaces. The seminal work of Cohen et al. on equivariant  G-CNNs on homogeneous spaces characterized the representation theory of such layers in the linear setting, finding that they are given by convolutions with kernels satisfying so-called steerability constraints. Motivated by the empirical success of non-linear layers, such as self-attention or input dependent kernels, we set out to generalize these insights to the non-linear setting. We derive generalized steerability constraints that any such layer needs to satisfy and prove the universality of our construction. The insights gained into the symmetry-constrained functional dependence of equivariant operators on feature maps and group elements informs the design of future equivariant neural network layers. We demonstrate how several common equivariant network architectures -  G-CNNs, implicit steerable kernel networks, conventional and relative position embedded attention based transformers, and LieTransformers - may be derived from our framework.

Författare

Elias Nyholm

Chalmers, Matematiska vetenskaper, Algebra och geometri

Oscar Carlsson

Chalmers, Matematiska vetenskaper, Algebra och geometri

Maurice Weiler

Massachusetts Institute of Technology (MIT)

Daniel Persson

Chalmers, Matematiska vetenskaper, Algebra och geometri

Fundament

Grundläggande vetenskaper

Ämneskategorier (SSIF 2025)

Geometri

Artificiell intelligens

DOI

10.48550/arXiv.2504.20974

Mer information

Senast uppdaterat

2025-09-03