首页 | 本学科首页   官方微博 | 高级检索  
     

知识感知的预训练语言模型综述
引用本文:李瑜泽,栾馨,柯尊旺,李哲,吾守尔·斯拉木.知识感知的预训练语言模型综述[J].计算机工程,2021,47(9):18-33.
作者姓名:李瑜泽  栾馨  柯尊旺  李哲  吾守尔·斯拉木
作者单位:1. 新疆大学 信息科学与工程学院, 乌鲁木齐 830046;2. 新疆多语种信息技术实验室, 乌鲁木齐 830046;3. 新疆多语种信息技术研究中心, 乌鲁木齐 830046;4. 新疆大学 软件学院, 乌鲁木齐 830046
基金项目:国家重点研发计划(2017YFC0820702-3)。
摘    要:随着自然语言处理(NLP)领域中预训练技术的快速发展,将外部知识引入到预训练语言模型的知识驱动方法在NLP任务中表现优异,知识表示学习和预训练技术为知识融合的预训练方法提供了理论依据。概述目前经典预训练方法的相关研究成果,分析在新兴预训练技术支持下具有代表性的知识感知的预训练语言模型,分别介绍引入不同外部知识的预训练语言模型,并结合相关实验数据评估知识感知的预训练语言模型在NLP各个下游任务中的性能表现。在此基础上,分析当前预训练语言模型发展过程中所面临的问题和挑战,并对领域发展前景进行展望。

关 键 词:自然语言处理  知识表征  语义知识  预训练  语言模型  
收稿时间:2021-02-05
修稿时间:2021-03-31

Survey of Knowledge-Aware Pre-Trained Language Models
LI Yuze,LUAN Xin,KE Zunwang,LI Zhe,Wushour Silamu.Survey of Knowledge-Aware Pre-Trained Language Models[J].Computer Engineering,2021,47(9):18-33.
Authors:LI Yuze  LUAN Xin  KE Zunwang  LI Zhe  Wushour Silamu
Affiliation:1. College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China;2. Xinjiang Laboratory of Multi-Language Information Technology, Urumqi 830046, China;3. Xinjiang Multi-Language Information Technology Research Center, Urumqi 830046, China;4. School of Software, Xinjiang University, Urumqi 830046, China
Abstract:In the field of Natural Language Processing(NLP), the recent years has witnessed a rapid development in pre-training technology, and the knowledge-driven method that injects external knowledge into a pre-trained language model performs excellently in NLP tasks.The techniques of knowledge representation learning and pre-training provide theoretical foundation for the knowledge-based pre-training method.This paper briefly introduces the development of the classical pre-trained methods.Then it analyzes the representative knowledge-aware pre-trained language models supported by new pre-training technology.According to the types of external knowledge, this paper introduces the pre-trained language models injected with different external knowledge.Based on relevant experimental data, it subsequently evaluates the performance of the knowledge-aware pre-trained language models in various downstream tasks of NLP.On this basis, the paper analyzes the problems and challenges faced by the developing pre-trained language models, and discusses the development trends of this field.
Keywords:Natural Language Processing(NLP)  knowledge representation  semantic knowledge  pre-training  language model  
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机工程》浏览原始摘要信息
点击此处可从《计算机工程》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号