首页 | 本学科首页   官方微博 | 高级检索  
     

融合双向依存自注意力机制的神经机器翻译
引用本文:李治瑾,赖华,文永华,高盛祥.融合双向依存自注意力机制的神经机器翻译[J].计算机应用,2022,42(12):3679-3685.
作者姓名:李治瑾  赖华  文永华  高盛祥
作者单位:昆明理工大学 信息工程与自动化学院,昆明 650504
云南省人工智能重点实验室(昆明理工大学),昆明 650504
基金项目:国家自然科学基金资助项目(61732005);云南省重大科技专项(202002AD080001?5);云南省高新技术产业专项(201606)
摘    要:针对神经机器翻译中资源稀缺的问题,提出了一种基于双向依存自注意力机制(Bi-Dependency)的依存句法知识融合方法。首先,利用外部解析器对源句子解析得到依存解析数据;然后,将依存解析数据转化为父词位置向量和子词权重矩阵;最后,将依存知识融合到Transformer编码器的多头注意力机制上。利用Bi-Dependency,翻译模型可以同时对父词到子词、子词到父词两个方向的依存信息进行关注。双向翻译的实验结果表明,与Transformer模型相比,在富资源情况下,所提方法在汉-泰翻译上的BLEU值分别提升了1.07和0.86,在汉-英翻译上的BLEU值分别提升了0.79和0.68;在低资源情况下,所提方法在汉-泰翻译上的BLEU值分别提升了0.51和1.06,在汉-英翻译上的BLEU值分别提升了1.04和0.40。可见Bi-Dependency为模型提供了更丰富的依存信息,能够有效提升翻译性能。

关 键 词:神经机器翻译  双向依存注意力  多头注意力  父词  子词  
收稿时间:2021-10-22
修稿时间:2022-01-06

Neural machine translation integrating bidirectional-dependency self-attention mechanism
Zhijin LI,Hua LAI,Yonghua WEN,Shengxiang GAO.Neural machine translation integrating bidirectional-dependency self-attention mechanism[J].journal of Computer Applications,2022,42(12):3679-3685.
Authors:Zhijin LI  Hua LAI  Yonghua WEN  Shengxiang GAO
Affiliation:Faculty of Information Engineering and Automation,Kunming University of Science and Technology,Kunming Yunnan 650504,China
Yunnan Key Laboratory of Artificial Intelligence (Kunming University of Science and Technology),Kunming Yunnan 650504,China
Abstract:Aiming at the problem of resource scarcity in neural machine translation, a method for fusion of dependency syntactic knowledge based on a Bidirectional-Dependency self-attention mechanism (Bi-Dependency) was proposed. Firstly, an external parser was used to parse the source sentence to obtain dependency parsing data. Then, the dependency parsing data was transformed into the position vector of the parent word and the weight matrix of the child word. Finally, the dependency knowledge was integrated into the multi-head attention mechanism of the Transformer encoder. By using Bi-Dependency, the translation model was able to simultaneously pay attention to the dependency information in both directions: the parent word to the child word and the child word to the parent word. Experimental results of bi-directional translation show that compared with the Transformer model, in the case of rich resources, the proposed method has the BLEU (BiLingual Evaluation Understudy) value on Chinese-Thai translation improved by 1.07 and 0.86 respectively, and the BLEU value on Chinese-English translation improved by 0.79 and 0.68 respectively; in the case of low resources, the proposed model has the BLEU value increased by 0.51 and 1.06 respectively on Chinese-Thai translation, and the BLEU value increased by 1.04 and 0.40 respectively on Chinese-English translation. It can be seen that Bi-Dependency provides the model with richer dependence information, which can effectively improve the translation performance.
Keywords:neural machine translation  bidirectional-dependency attention  multi-head attention  parent word  child word  
点击此处可从《计算机应用》浏览原始摘要信息
点击此处可从《计算机应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号