Equivariant neural networks encode symmetry as an inductive bias and have achieved strong empirical performance in wide domains. However, their expressive power remains not well understood. Focusing on 2-layer ReLU networks, this paper investigates the impact of equivariance constraints on the expressivity of equivariant and layer-wise equivariant networks. By examining the boundary hyperplanes and the channel vectors of ReLU networks, we construct an example showing that equivariance constraints could strictly limit expressive power. However, we demonstrate that this drawback can be compensated via enlarging the model size. Furthermore, we show that despite a larger model size, the resulting architecture could still correspond to a hypothesis space with lower complexity, implying superior generalizability for equivariant networks.
翻译:等变神经网络将对称性编码为归纳偏置,在广泛领域中取得了显著的实证性能。然而,其表达能力仍未得到充分理解。本文聚焦于两层ReLU网络,研究了等变约束对等变网络与逐层等变网络表达能力的影响。通过分析ReLU网络的边界超平面与通道向量,我们构建了一个实例,表明等变约束可能严格限制表达能力。然而,我们证明这一局限性可以通过扩大模型规模得到补偿。此外,我们发现即使模型规模更大,所得架构仍可能对应假设空间复杂度更低,这意味着等变网络具有更优的泛化能力。