首页 | 本学科首页   官方微博 | 高级检索  
     

TP-FER:基于优化卷积神经网络的三通道人脸表情识别方法
引用本文:高静文,蔡永香,何宗宜.TP-FER:基于优化卷积神经网络的三通道人脸表情识别方法[J].计算机应用研究,2021,38(7):2213-2219.
作者姓名:高静文  蔡永香  何宗宜
作者单位:地理信息工程国家重点实验室,西安710054;长江大学 地球科学学院,武汉430100;长江大学 地球科学学院,武汉430100
基金项目:地理信息工程国家重点实验室开放基金资助项目(SKLGIE2017-M-4-6);国家自然科学基金青年基金资助项目(41701537)
摘    要:针对人脸五官在表达不同情绪时所起的作用不同,利用单一的卷积神经网络对人脸面部特征进行特征提取和表情识别可能会导致提取表情关键特征信息时聚焦性不够,而仅对眼部或者嘴部等重点部位进行特征提取,又有可能造成特征提取不够充分的问题,提出了一种基于优化卷积神经网络的三通道人脸表情识别方法TP-FER(tri-path networks for facial expression recognition).该方法基于构建的卷积神经网络训练,采用三个输入渠道,分别聚焦面部、眼部和嘴部区域进行特征提取和表情判别,最后采用基于决策层的融合技术将三个渠道的识别结果进行相对多数投票决策,获取整体最优识别率.将此方法应用于JCK+数据集和自建数据集上进行了实验判别分析,结果表明该方法在两个数据集上均提高了整体表情识别率.该方法既考虑了脸部整体特征的提取,又兼顾了某些表情主要聚焦在眼部、嘴部表达的特性,相互辅助,整体提高了表情的识别率;该方法也能对神经心理学研究提供数据支持.

关 键 词:人脸表情识别  卷积神经网络  聚焦  驾驶情绪
收稿时间:2020/7/10 0:00:00
修稿时间:2021/6/17 0:00:00

TP-FER:facial expression recognition method of tri-path networks based on optimal convolutional neural network
Gao Jingwen,Cai Yongxiang and He Zongyi.TP-FER:facial expression recognition method of tri-path networks based on optimal convolutional neural network[J].Application Research of Computers,2021,38(7):2213-2219.
Authors:Gao Jingwen  Cai Yongxiang and He Zongyi
Affiliation:School of Geosciences,Yangtze University,Wuhan Hubei,,
Abstract:Facial features play different functions in expressing different emotions, the only use of convolutional neural network for feature extraction and expression recognition of facial features may not lead to enough focus on the key feature information of facial expressions. However, the feature information only extracting from the key parts such as eyes or mouth may not be sufficient. This paper proposed a facial expression recognition method of tri-path networks based on the optimal convolutional neural network called TP-FER. Based on the convolutional neural network, this method used three input channels to focus on the face, eyes and mouth regions respectively for feature extraction and expression discrimination. Then it used the fusion technology based on the decision layer to make the final decision. At last'' it adopted relative majority voting method based on the recognition results of the three channels to obtain the overall optimal recognition rate. This paper conducted several experiments on JCK+dataset and self-built dataset with this method. The results show that the overall expression recognition rate is improved in both datasets. This method can make full use of the different channel characteristics that focus on not only the key features such as eyes and mouth but also the comprehensive information of the whole face, and improve the expression recognition rate effectively. In addition, it can also provide data support for neuropsychological related research.
Keywords:facial expression recognition  convolutional neural network  focus  driving emotion
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机应用研究》浏览原始摘要信息
点击此处可从《计算机应用研究》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号