Word order is a significant distinctive feature to differentiate languages. In this paper, we investigate cross-lingual transfer and posit that an order-agnostic model will perform better when transferring to distant foreign languages. To test our hypothesis, we train dependency parsers on an English corpus and evaluate their transfer performance on 30 other languages. Specifically, we compare encoders and decoders based on Recurrent Neural Networks (RNNs) and modified self-attentive architectures. The former rely on sequential information while the latter are more flexible at modeling token order. Detailed analysis shows that RNN-based architectures transfer well to languages that are close to English, while self-attentive models have better overall cross-lingual transferability and perform especially well on distant languages.
翻译:单词顺序是区分语言的重要特征。 在本文中, 我们调查跨语言传输, 并假设在向遥远的外语转移时, 命令不可知性模式效果会更好 。 为了测试我们的假设, 我们培训英国文的依赖分析员, 并评估其他30种语言的转移表现 。 具体地说, 我们比较基于经常性神经网络( RNN) 和 修改过的自适应结构的编码器和解码器。 前者依赖于相继信息, 而后者在模拟象征性命令上则比较灵活 。 详细分析显示, 以 RNN 为基础的结构向接近英语的语言转移得很好, 而自强化模型在整体上更便于跨语言的传输, 并且对遥远的语言表现得特别好 。