首页 | 本学科首页   官方微博 | 高级检索  
     


Video-based emotion recognition in the wild using deep transfer learning and score fusion
Affiliation:1. Department of Computer Engineering, Çorlu Faculty of Engineering, Namık Kemal University, 59860 Çorlu, Tekirdağ, Turkey;2. Department of Computer Engineering, Boğaziçi University, 34342 Bebek, İstanbul, Turkey;1. School of Automation, China University of Geosciences, Wuhan 430074, China;2. Hubei Key Laboratory of Advanced Control and Intelligent Automation for Complex Systems, Wuhan 430074, China;1. Department of Computer Engineering Namık Kemal University, Çorlu 59860, Tekirdağ, Turkey;2. St. Petersburg Institute for Informatics and Automation of the Russian Academy of Sciences (SPIIRAS), 199178, St. Petersburg, Russia
Abstract:Multimodal recognition of affective states is a difficult problem, unless the recording conditions are carefully controlled. For recognition “in the wild”, large variances in face pose and illumination, cluttered backgrounds, occlusions, audio and video noise, as well as issues with subtle cues of expression are some of the issues to target. In this paper, we describe a multimodal approach for video-based emotion recognition in the wild. We propose using summarizing functionals of complementary visual descriptors for video modeling. These features include deep convolutional neural network (CNN) based features obtained via transfer learning, for which we illustrate the importance of flexible registration and fine-tuning. Our approach combines audio and visual features with least squares regression based classifiers and weighted score level fusion. We report state-of-the-art results on the EmotiW Challenge for “in the wild” facial expression recognition. Our approach scales to other problems, and ranked top in the ChaLearn-LAP First Impressions Challenge 2016 from video clips collected in the wild.
Keywords:
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号