首页 | 本学科首页   官方微博 | 高级检索  
     

基于ACNNC模型的中文分词方法
引用本文:张忠林,余炜,闫光辉,袁晨予. 基于ACNNC模型的中文分词方法[J]. 中文信息学报, 2022, 36(8): 12-19+28
作者姓名:张忠林  余炜  闫光辉  袁晨予
作者单位:兰州交通大学 电子与信息工程学院,甘肃 兰州 730070
基金项目:国家自然科学基金(61662043,62062049);甘肃省哲学社会科学规划项目(20YB056)
摘    要:目前,现有中文分词模型大多是基于循环神经网络的,其能够捕捉序列整体特征,但存在忽略了局部特征的问题。针对这种问题,该文综合了注意力机制、卷积神经网络和条件随机场,提出了注意力卷积神经网络条件随机场模型(Attention Convolutional Neural Network CRF, ACNNC)。其中,嵌入层训练词向量,自注意力层代替循环神经网络捕捉序列全局特征,卷积神经网络捕捉序列局部特征和位置特征,经融合层的特征输入条件随机场进行解码。实验证明该文提出的模型在BACKOFF 2005测试集上有更好的分词效果,并在PKU、MSR、CITYU和AS上取得了96.2%、96.4%、96.1%和95.8%的F1值。

关 键 词:中文分词  深度学习  注意力机制

Chinese Word Segmentation Based on ACNNC Model
ZHANG Zhonglin,YU Wei,YAN Guanghui,YUAN Chenyu. Chinese Word Segmentation Based on ACNNC Model[J]. Journal of Chinese Information Processing, 2022, 36(8): 12-19+28
Authors:ZHANG Zhonglin  YU Wei  YAN Guanghui  YUAN Chenyu
Affiliation:School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou, Gansu 730070, China
Abstract:At present, most of the existing Chinese word segmentation models are based on recurrent neural networks, which can capture the overall features of the sequence while ignoring local features. This paper combines the attention mechanism, convolutional neural network and conditional random fields, and proposes Attention Convolutional Neural Network CRF (ACNNC). The self-attention layer replaces the recurrent neural network to capture the global features of the sequence, and the convolutional neural network captures location features of the sequence. The features are combined in the fusion layer and then input into conditional random fields for decoding. The experimental results on BACKOFF 2005 show that the model proposed achieves 96.2%, 96.4%, 96.1% and 95.8% F1 values on PKU, MSR, CITYU and AS test set, respectively.
Keywords:Chinese word segmentation    deep learning    attention mechanism  
点击此处可从《中文信息学报》浏览原始摘要信息
点击此处可从《中文信息学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号