Published January 17, 2024 | Version v1
Conference proceeding Open

A Characterization Theorem for Equivariant Networks with Point-wise Activations

  • 1. ROR icon Fondazione Bruno Kessler
  • 2. University of Oxford

Description

Equivariant neural networks have shown improved performance, expressiveness and sample complexity on symmetrical domains. But for some specific symmetries, representations, and choice of coordinates, the most common point-wise activations, such as ReLU, are not equivariant, hence they cannot be employed in the design of equivariant neural networks. The theorem we present in this paper describes all possibile combinations of representations, choice of coordinates and point-wise activations to obtain an equivariant layer, generalizing and strengthening existing characterizations. Notable cases of practical relevance are discussed as corollaries. Indeed, we prove that rotation-equivariant networks can only be invariant, as it happens for any network which is equivariant with respect to connected compact groups. Then, we discuss implications of our findings when applied to important instances of equivariant networks. First, we completely characterize permutation equivariant networks such as Invariant Graph Networks with point-wise nonlinearities and their geometric counterparts, highlighting a plethora of models whose expressive power and performance are still unknown. Second, we show that feature spaces of disentangled steerable convolutional neural networks are trivial representations.

Files

6118_a_characterization_theorem_for.pdf

Files (342.9 kB)

Name Size Download all
md5:cdfda8a1eacb62a895f7ea4c6622f890
342.9 kB Preview Download

Additional details

Funding

ELIAS – European Lighthouse of AI for Sustainability 101120237
European Commission