首页 | 本学科首页   官方微博 | 高级检索  
     


Cross-dataset emotion recognition from facial expressions through convolutional neural networks
Affiliation:1. Department of Information Engineering and Computer Science, Feng Chia University, Taichung 40724, Taiwan;2. Department of Computer Science, University of Warwick, Coventry CV4 7AL, UK;3. Department of Computer Science and Information Engineering, National Chin-Yi University of Technology, Taichung 41170, Taiwan;1. Electronics Engineering, Bharati Vidyapeeth (Deemed to be University) College of Engineering, Pune 411043, Maharashtra, India;2. E&TC Engineering, TSSM''s Padmabhooshan Vasantdada Patil Institute of Technology, Pune 411043, Maharashtra, India;3. E&TC Engineering, Vishwakarma Institute of Technology, Pune 411037, Maharashtra, India;1. B3 Building, School of Computer Science and Engineering, South China University of Technology, Guangzhou Higher Education Mega Center, Panyu District 510006, Guangzhou, PR China;2. B7 Building, School of Software Engineering, South China University of Technology, Guangzhou Higher Education Mega Center, Panyu District 510006, Guangzhou, PR China
Abstract:The face is the window to the soul. This is what the 19th-century French doctor Duchenne de Boulogne thought. Using electric shocks to stimulate muscular contractions and induce bizarre-looking expressions, he wanted to understand how muscles produce facial expressions and reveal the most hidden human emotions. Two centuries later, this research field remains very active. We see automatic systems for recognizing emotion and facial expression being applied in medicine, security and surveillance systems, advertising and marketing, among others. However, there are still fundamental questions that scientists are trying to answer when analyzing a person’s emotional state from their facial expressions. Is it possible to reliably infer someone’s internal state based only on their facial muscles’ movements? Is there a universal facial setting to express basic emotions such as anger, disgust, fear, happiness, sadness, and surprise? In this research, we seek to address some of these questions through convolutional neural networks. Unlike most studies in the prior art, we are particularly interested in examining whether characteristics learned from one group of people can be generalized to predict another’s emotions successfully. In this sense, we adopt a cross-dataset evaluation protocol to assess the performance of the proposed methods. Our baseline is a custom-tailored model initially used in face recognition to categorize emotion. By applying data visualization techniques, we improve our baseline model, deriving two other methods. The first method aims to direct the network’s attention to regions of the face considered important in the literature but ignored by the baseline model, using patches to hide random parts of the facial image so that the network can learn discriminative characteristics in different regions. The second method explores a loss function that generates data representations in high-dimensional spaces so that examples of the same emotion class are close and examples of different classes are distant. Finally, we investigate the complementarity between these two methods, proposing a late-fusion technique that combines their outputs through the multiplication of probabilities. We compare our results to an extensive list of works evaluated in the same adopted datasets. In all of them, when compared to works that followed an intra-dataset protocol, our methods present competitive numbers. Under a cross-dataset protocol, we achieve state-of-the-art results, outperforming even commercial off-the-shelf solutions from well-known tech companies.
Keywords:Emotion recognition  Facial analysis  Cross-dataset evaluation  Deep learning
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号