Entity Linking (EL) has traditionally relied on large annotated datasets and extensive model fine-tuning. While recent few-shot methods leverage large language models (LLMs) through prompting to reduce training requirements, they often suffer from inefficiencies due to expensive LLM-based reasoning. ARTER (Adaptive Routing and Targeted Entity Reasoning) presents a structured pipeline that achieves high performance without deep fine-tuning by strategically combining candidate generation, context-based scoring, adaptive routing, and selective reasoning. ARTER computes a small set of complementary signals(both embedding and LLM-based) over the retrieved candidates to categorize contextual mentions into easy and hard cases. The cases are then handled by a low-computational entity linker (e.g. ReFinED) and more expensive targeted LLM-based reasoning respectively. On standard benchmarks, ARTER outperforms ReFinED by up to +4.47%, with an average gain of +2.53% on 5 out of 6 datasets, and performs comparably to pipelines using LLM-based reasoning for all mentions, while being as twice as efficient in terms of the number of LLM tokens.
翻译:实体链接传统上依赖于大规模标注数据集和广泛的模型微调。尽管近期少样本方法通过提示利用大语言模型以减少训练需求,但它们常因基于大语言模型的昂贵推理而效率低下。ARTER(自适应路由与定向实体推理)提出了一种结构化流程,通过策略性地结合候选生成、基于上下文的评分、自适应路由和选择性推理,在不进行深度微调的情况下实现了高性能。ARTER在检索到的候选实体上计算一组小型互补信号(包括嵌入和大语言模型信号),将上下文提及分类为简单和困难案例。这些案例随后分别由低计算成本的实体链接器(如ReFinED)和更昂贵的基于大语言模型的定向推理处理。在标准基准测试中,ARTER在6个数据集中的5个上平均提升+2.53%,最高超越ReFinED达+4.47%,且与对所有提及均使用大语言模型推理的流程性能相当,同时在大语言模型令牌使用数量上效率提升一倍。