Context-aware translation can be achieved by processing a concatenation of consecutive sentences with the standard translation approach. This paper investigates the intuitive idea of adopting segment embeddings for this task to help the Transformer discern the position of each sentence in the concatenation sequence. We compare various segment embeddings and propose novel methods to encode sentence position into token representations, showing that they do not benefit the vanilla concatenation approach except in a specific setting.
翻译:通过采用标准翻译方法处理连续几句的顺序,可以实现背景意识翻译。本文件探讨为这项任务采用分层嵌入的直觉想法,以帮助变换器辨别每个句子在连接序列中的位置。我们比较了各段嵌入,并提出了将句子位置编码为象征性表达的新方法,表明它们除了在特定环境下,不会有利于香草凝聚方法。