Inductive link prediction is emerging as a key paradigm for real-world knowledge graphs (KGs), where new entities frequently appear and models must generalize to them without retraining. Predicting links in a KG faces the challenge of guessing previously unseen entities by leveraging generalizable node features such as subgraph structure, type annotations, and ontological constraints. However, explicit type information is often lacking or incomplete. Even when available, type information in most KGs is often coarse-grained, sparse, and prone to errors due to human annotation. In this work, we explore the potential of pre-trained language models (PLMs) to enrich node representations with implicit type signals. We introduce TyleR, a Type-less yet type-awaRe approach for subgraph-based inductive link prediction that leverages PLMs for semantic enrichment. Experiments on standard benchmarks demonstrate that TyleR outperforms state-of-the-art baselines in scenarios with scarce type annotations and sparse graph connectivity. To ensure reproducibility, we share our code at https://github.com/sisinflab/tyler .
翻译:归纳式链接预测正成为现实世界知识图谱(KGs)的关键范式,其中新实体频繁出现,模型必须在不重新训练的情况下泛化到这些新实体。在知识图谱中预测链接面临着通过利用可泛化的节点特征(如子图结构、类型标注和本体约束)来猜测先前未见实体的挑战。然而,显式的类型信息常常缺失或不完整。即使可用,大多数知识图谱中的类型信息也通常是粗粒度的、稀疏的,并且由于人工标注容易出错。在这项工作中,我们探索了预训练语言模型(PLMs)利用隐式类型信号丰富节点表示的潜力。我们提出了TyleR,一种无类型但类型感知的基于子图的归纳式链接预测方法,它利用PLMs进行语义增强。在标准基准测试上的实验表明,在类型标注稀缺和图连接稀疏的场景下,TyleR优于最先进的基线方法。为确保可复现性,我们在 https://github.com/sisinflab/tyler 分享了代码。