Graph machine learning architectures are typically tailored to specific tasks on specific datasets, which hinders their broader applicability. This has led to a new quest in graph machine learning: how to build graph foundation models capable of generalizing across arbitrary graphs and features? In this work, we present a recipe for designing graph foundation models for node-level tasks from first principles. The key ingredient underpinning our study is a systematic investigation of the symmetries that a graph foundation model must respect. In a nutshell, we argue that label permutation-equivariance alongside feature permutation-invariance are necessary in addition to the common node permutation-equivariance on each local neighborhood of the graph. To this end, we first characterize the space of linear transformations that are equivariant to permutations of nodes and labels, and invariant to permutations of features. We then prove that the resulting network is a universal approximator on multisets that respect the aforementioned symmetries. Our recipe uses such layers on the multiset of features induced by the local neighborhood of the graph to obtain a class of graph foundation models for node property prediction. We validate our approach through extensive experiments on 29 real-world node classification datasets, demonstrating both strong zero-shot empirical performance and consistent improvement as the number of training graphs increases.
翻译:图机器学习架构通常针对特定数据集上的特定任务进行定制,这限制了其更广泛的适用性。这引发了图机器学习领域的新探索:如何构建能够泛化至任意图和特征的图基础模型?在本研究中,我们提出了从第一性原理设计节点级任务图基础模型的方法论。本研究的核心基础是对图基础模型必须遵循的对称性进行的系统性研究。简而言之,我们认为除了图每个局部邻域上常见的节点置换等变性外,标签置换等变性与特征置换不变性同样不可或缺。为此,我们首先刻画了满足节点与标签置换等变性、同时保持特征置换不变性的线性变换空间。随后我们证明,所得到的网络在遵循上述对称性的多重集上具有通用逼近能力。我们的方法通过在图的局部邻域诱导的特征多重集上应用此类层,获得了一类用于节点属性预测的图基础模型。我们在29个真实世界节点分类数据集上通过大量实验验证了该方法,既展示了强大的零样本实证性能,也证明了随着训练图数量增加而持续提升的稳定性。