首页 | 本学科首页   官方微博 | 高级检索  
     

基于ELMO的低资源神经机器翻译
引用本文:王浩畅,孙孟冉,赵铁军.基于ELMO的低资源神经机器翻译[J].计算机与现代化,2021,0(7):38-42.
作者姓名:王浩畅  孙孟冉  赵铁军
作者单位:东北石油大学计算机与信息技术学院,黑龙江 大庆 163318;哈尔滨工业大学计算机科学与技术学院,黑龙江 哈尔滨 150001
基金项目:国家自然科学基金资助项目(61402099, 61702093)
摘    要:低资源神经机器翻译的研究难点是缺乏大量的平行语料来给模型进行训练。随着预训练模型的发展,并且在各大自然语言处理任务中均取得很大的提升,本文提出一种融合ELMO预训练模型的神经机器翻译模型来解决低资源神经机器翻译问题。本文模型在土耳其语-英语低资源翻译任务上相比于反向翻译提升超过0.7个BLEU,在罗马尼亚语-英语翻译任务上提升超过0.8个BLEU。此外,在模拟的中-英、法-英、德-英、西-英这4组低资源翻译任务上相比于传统神经机器翻译模型分别提升2.3、3.2、2.6、3.2个BLEU。实验表明使用融合ELMO的模型来解决低资源神经机器翻译问题是有效的。

关 键 词:低资源  平行语料  预训练模型  神经机器  翻译模型  
收稿时间:2021-08-02

Low-resource Neural Machine Translation Based on ELMO
WANG Hao-chang,SUN Meng-ran,ZHAO Tie-jun.Low-resource Neural Machine Translation Based on ELMO[J].Computer and Modernization,2021,0(7):38-42.
Authors:WANG Hao-chang  SUN Meng-ran  ZHAO Tie-jun
Abstract:The difficulty in low-resource neural machine translation is lack of numerous parallel corpus to train the model. With the development of the pre-training model, it has made great improvements in various natural language processing tasks. In this paper, a neural machine translation model combining ELMO is proposed to solve the problem of low-resource neural machine translation. There are more than 0.7 BLEU improvements in the Turkish-English low-resource translation task compared to the back translation, and more than 0.8 BLEU improvements in the Romanian-English translation task. In addition, compared with the traditional neural machine translation model, the simulated low-resource translation tasks of Chinese-English, French-English, German-English and Spanish-English increase by 2.3, 3.2, 2.6 and 3.2 BLEU respectively. The experimental results show that the ELMO model is effective for low-resource neural machine translation.
Keywords:low-resource  parallel corpus  pre-training model  neural machine  translation model  
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机与现代化》浏览原始摘要信息
点击此处可从《计算机与现代化》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号