Does equivariance make better models?

Aug 1, 2024, 11:40 AM
20m
Palazzo Ducale (Genova, Italy)

Palazzo Ducale

Genova, Italy

Talk Novel Techniques Novel Techniques

Speaker

Dr Alexander Bogatskiy

Description

Abstract: The current best-performing networks in many ML for particle physics tasks are either custom-built Lorentz-equivariant architectures or more generic large transformer models. A major unanswered question is whether the high performance of equivariant architectures is in fact due to their equivariance. We design a study to isolate and investigate effects of equivariance on network performance. A particular equivariant model, PELICAN, has its symmetry broken down with no to minimal architectural changes via two methods. First, equivariance is broken “explicitly" by supplying model inputs that are equivariant under proper subgroups of the Lorentz group. In the second method, it is broken “implicitly" by adding spurious particles which encode laboratory-frame geometry. We compare its performance on common benchmark tasks in the equivariant and non-equivariant regimes.

Primary author

Presentation materials