Poster
Equivariant Neural Tangent Kernels
Philipp Misof · Pan Kessel · Jan Gerken
West Exhibition Hall B2-B3 #W-718
A significant proportion of machine learning problems are subject to inherent symmetries, e.g. translation or rotation symmetry in image classification. This underlying property can either be learned by means of data augmentation or enforced through the model structure itself, so-called equivariant networks. However, the training dynamics of the latter are not understood well, rendering a systematic comparison between those approaches difficult.In this work, we extend a mathematical tool, called the neural tangent kernel (NTK) to equivariant networks. In the regime of wide hidden layers, it allows for an analytic solution of the training dynamics and has already been applied with great success on simpler models. We use our extension to find an explicit connection between data augmentation and equivariance: For a particular class of conventional networks trained with data augmentation, we find corresponding equivariant networks that share the same expected training dynamics in the limit of infinitely wide hidden layers. We further implement the general analytic relations we found for rotation and translation symmetries.The presented framework and its implications contribute to the ongoing debate on whether symmetry should be enforced by construction or learned from data by finding an explicit correspondence between those two approaches.