Equivariant Neural Tangent Kernels
Paper i proceeding, 2025

Little is known about the training dynamics of equivariant neural networks, in particular how it compares to data augmented training of their non-equivariant counterparts. Recently, neural tangent kernels (NTKs) have emerged as a powerful tool to analytically study the training dynamics of wide neural networks. In this work, we take an important step towards a theoretical understanding of training dynamics of equivariant models by deriving neural tangent kernels for a broad class of equivariant architectures based on group convolutions. As a demonstration of the capabilities of our framework, we show an interesting relationship between data augmentation and group convolutional networks. Specifically, we prove that they share the same expected prediction over initializations at all training times and even off the data manifold. In this sense, they have the same training dynamics. We demonstrate in numerical experiments that this still holds approximately for finite-width ensembles. By implementing equivariant NTKs for roto-translations in the plane (G = Cn ⋉ R2) and 3d rotations (G = SO(3)), we show that equivariant NTKs outperform their non-equivariant counterparts as kernel predictors for histological image classification and quantum mechanical property prediction.

Författare

Philipp Misof

Göteborgs universitet

Chalmers, Matematiska vetenskaper, Algebra och geometri

Pan Kessel

F. Hoffmann-La Roche AG

Jan Gerken

Chalmers, Matematiska vetenskaper, Algebra och geometri

Göteborgs universitet

Proceedings of the 38th Conference on Uncertainty in Artificial Intelligence, UAI 2022

26403498 (eISSN)

Vol. 267 44470-44503

42nd International Conference on Machine Learning, ICML 2025
Vancouver, Canada,

Ämneskategorier (SSIF 2025)

Datavetenskap (datalogi)

Infrastruktur

C3SE (-2020, Chalmers Centre for Computational Science and Engineering)

Mer information

Senast uppdaterat

2025-12-17