Pre-trained language models (PLMs) have made remarkable progress in table-to-text generation tasks. However, the topological gap between tabular data and text and the lack of domain-specific knowledge make it difficult for PLMs to produce faithful text, especially in real-world applications with limited resources. In this paper, we mitigate the above challenges by introducing a novel augmentation method: Prompt-based Adapter (PA), which targets table-to-text generation under few-shot conditions. The core insight design of the PA is to inject prompt templates for augmenting domain-specific knowledge and table-related representations into the model for bridging the structural gap between tabular data and descriptions through adapters. Such prompt-based knowledge augmentation method brings at least two benefits: (1) enables us to fully use the large amounts of unlabelled domain-specific knowledge, which can alleviate the PLMs' inherent shortcomings of lacking domain knowledge; (2) allows us to design different types of tasks supporting the generative challenge. Extensive experiments and analyses are conducted on three open-domain few-shot NLG datasets: Humans, Books, and Songs. Compared to previous state-of-the-art approaches, our model achieves superior performance in terms of both fluency and accuracy as judged by human and automatic evaluations.
翻译:事先培训的语言模型(PLM)在制表-制表任务方面取得了显著进展,然而,表格数据和文本之间的地形差距以及缺乏具体领域的知识使PLM难以产生忠实的文本,特别是在资源有限的现实应用中。在本文件中,我们通过采用新的增强方法减轻上述挑战:即快速调适器(PA),它针对在少见条件下的表格-文本生成;PA的核心洞察设计是迅速为通过适应者弥合表格数据与描述之间的结构性差距的模型输入扩大特定领域知识和表格相关表述的模板。这种快速知识增强方法至少带来两个好处:(1) 使我们能够充分利用大量无标签的特定领域知识,这可以减轻PLMS缺乏领域知识的固有缺陷;(2) 使我们能够设计不同种类的任务,支持归正挑战。 广泛实验和分析针对三种未露面的NLG数据集:人类、书籍和歌曲。与我们以往的高级和自动评估方法相比,在人类业绩评估方面实现了高超度和高度。</s>