首页 | 本学科首页   官方微博 | 高级检索  
     

文本词向量与预训练语言模型研究
引用本文:徐菲菲,冯东升.文本词向量与预训练语言模型研究[J].上海电力学院学报,2020,36(4):320-328.
作者姓名:徐菲菲  冯东升
作者单位:上海电力大学 计算机科学与技术学院
基金项目:上海市自然科学基金(19ZR1420800)。
摘    要:介绍了文本词向量及预训练语言模型的发展体系,系统整理并分析了其中重点方法的思想特点。首先,阐述了传统的文本词向量表征方法及基于语言模型的文本表征方法;然后,详述了预训练语言模型方法的研究进展,包括动态词向量的表征方法和基于Transformer架构的预训练模型;最后,指出了未来探究多模态间更有效的融合方式和迁移学习将成为该领域的发展趋势。

关 键 词:文本信息处理  词向量  预训练语言模型  Transformer架构
收稿时间:2020/3/18 0:00:00

A Survey of Research on Word Vectors and Pretrained Language Models
XU Feifei,FENG Dongsheng.A Survey of Research on Word Vectors and Pretrained Language Models[J].Journal of Shanghai University of Electric Power,2020,36(4):320-328.
Authors:XU Feifei  FENG Dongsheng
Affiliation:School of Computer Science and Technology, Shanghai University of Electric Power, Shanghai 200090, China
Abstract:This paper mainly introduces the development system of text word vectors and pre-trained language models,systematically organizes and analyzes the ideological characteristics of key methods.Firstly,we describe the traditional text word vector representation method and the language model-based text representation method,then we elaborate the research progress of the pre-trained language model method,including the dynamic word vector representation method and the Transformer architecture-based pre-training model.Finally,it is pointed out that in the future,exploring more effective fusion methods and transfer learning between multi-modalities will become a development trend in this field.
Keywords:text information processing  word vector  pre-trained language model  Transformer architecture
本文献已被 CNKI 等数据库收录!
点击此处可从《上海电力学院学报》浏览原始摘要信息
点击此处可从《上海电力学院学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号