Long-range transformer models have achieved encouraging results on long-context question answering (QA) tasks. Such tasks often require reasoning over a long document, and they benefit from identifying a set of evidence spans (e.g., sentences) that provide supporting evidence for addressing the question. In this work, we propose a novel method for equipping long-range transformers with an additional sequence-level objective for better identification of supporting evidence spans. We achieve this by proposing an additional contrastive supervision signal in finetuning, where the model is encouraged to explicitly discriminate supporting evidence sentences from negative ones by maximizing the question-evidence similarity. The proposed additional loss exhibits consistent improvements on three different strong long-context transformer models, across two challenging question answering benchmarks - HotpotQA and QAsper.
翻译:长程变压器模型在长文本答题(QA)任务方面取得了令人鼓舞的结果,这些任务往往需要长文件的推理,它们从确定一系列证据范围(如句子)中受益,为解决问题提供了支持性证据。在这项工作中,我们提出了一种新的方法,为长程变压器配备一个额外的序列级目标,以更好地识别支持性证据范围。我们通过在微调中提出另一个对比性监督信号来实现这一目标,鼓励该模型通过最大限度地扩大问题证据相似性,明确区分支持性证据判决与否定性判决。提议的额外损失表明,在两个具有挑战性的问题回答基准(HotpotQA和QAsper)之间,三种不同的强型长文本变压器模型不断改进。