Equivariant Neural Tangent Kernels
Paper in proceeding, 2025

Little is known about the training dynamics of equivariant neural networks, in particular how it compares to data augmented training of their non-equivariant counterparts. Recently, neural tangent kernels (NTKs) have emerged as a powerful tool to analytically study the training dynamics of wide neural networks. In this work, we take an important step towards a theoretical understanding of training dynamics of equivariant models by deriving neural tangent kernels for a broad class of equivariant architectures based on group convolutions. As a demonstration of the capabilities of our framework, we show an interesting relationship between data augmentation and group convolutional networks. Specifically, we prove that they share the same expected prediction over initializations at all training times and even off the data manifold. In this sense, they have the same training dynamics. We demonstrate in numerical experiments that this still holds approximately for finite-width ensembles. By implementing equivariant NTKs for roto-translations in the plane (G = Cn ⋉ R2) and 3d rotations (G = SO(3)), we show that equivariant NTKs outperform their non-equivariant counterparts as kernel predictors for histological image classification and quantum mechanical property prediction.

Author

Philipp Misof

University of Gothenburg

Chalmers, Mathematical Sciences, Algebra and geometry

Pan Kessel

F. Hoffmann-La Roche AG

Jan Gerken

Chalmers, Mathematical Sciences, Algebra and geometry

University of Gothenburg

Proceedings of the 38th Conference on Uncertainty in Artificial Intelligence, UAI 2022

26403498 (eISSN)

Vol. 267 44470-44503

42nd International Conference on Machine Learning, ICML 2025
Vancouver, Canada,

Subject Categories (SSIF 2025)

Computer Sciences

Infrastructure

C3SE (-2020, Chalmers Centre for Computational Science and Engineering)

More information

Latest update

12/17/2025