首页 | 本学科首页   官方微博 | 高级检索  
     

基于多视角学习的时序多模态情感分类研究
引用本文:陶全桧,安俊秀,戴宇睿,陈宏松,黄萍.基于多视角学习的时序多模态情感分类研究[J].计算机应用研究,2023,40(1):102-106.
作者姓名:陶全桧  安俊秀  戴宇睿  陈宏松  黄萍
作者单位:成都信息工程大学 软件工程学院,成都信息工程大学 软件工程学院,成都信息工程大学 软件工程学院,成都信息工程大学 软件工程学院,成都信息工程大学 管理学院
基金项目:国家自然科学基金资助项目(71673032);四川省社会科学高水平团队基金资助项目(2015Z177)
摘    要:针对多模态融合效果不佳,不能充分挖掘特定时间段,多视角关键情感信息的问题,提出了一种基于多视角的时序多模态情感分类模型,用于提取特定时间段,多视角下的关键情感信息。首先,对文本标题及文本内容两种视角下的数据进行低维空间词嵌入和序列表达,提取不同视角的多模态时序特征,对图片截取,水平镜像两种视角下的数据进行特征提取;其次,采用循环神经网络构建多模态数据的时序序列交互特征,增大互信息;最后,基于对比学习进行联合训练,完成情感分类。该模型在两个多模态情感分类基准数据集Yelp和Mutli-Zol上评估,准确度分别为73.92%、69.15%。综合实验表明,多视角的特定时间段多模态语句序列可提升模型性能。

关 键 词:情感分类  多模态  多视角  时序特征  对比学习
收稿时间:2022/6/22 0:00:00
修稿时间:2022/12/25 0:00:00

Temporal multimodal sentiment classification based on multi-view learning
Taoquanhui,Anjunxiu,Daiyurui,Chenhongsong and Huangping.Temporal multimodal sentiment classification based on multi-view learning[J].Application Research of Computers,2023,40(1):102-106.
Authors:Taoquanhui  Anjunxiu  Daiyurui  Chenhongsong and Huangping
Abstract:Aiming at the problem that the multi-modal fusion effect is not good, and the key emotional information from specific time periods and multiple perspectives cannot be fully mined, this paper proposed a time-series multi-modal emotional classification model based on multiple perspectives to extrout key information from multiple perspectives in a specific time period. Firstly, it performed low-dimensional spatial word embedding and sequence expression on the data from the two perspectives of text title and text content, extracted multi-modal time series features from different perspectives, and performed feature extraction on the data from two perspectives of image interception and horizontal mirroring. Secondly, it used the recurrent neural network to construct the time series interaction features of multi-modal data to increase mutual information. Finally, joint training was performed based on contrastive learning to complete sentiment classification. The model was evaluated on two multi-modal sentiment classification benchmark datasets Yelp and Mutli-Zol, with accuracies of 73.92% and 69.15%, respectively. Comprehensive experiments show that multi-view multi-modal sentence sequences in specific time periods can improve model performance.
Keywords:sentiment classification  multimodality  multiview  temporal features  contrastive learning
点击此处可从《计算机应用研究》浏览原始摘要信息
点击此处可从《计算机应用研究》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号