首页 | 本学科首页   官方微博 | 高级检索  
     


A review of affective computing: From unimodal analysis to multimodal fusion
Affiliation:1. School of Natural Sciences, University of Stirling, UK;2. Temasek Laboratories, Nanyang Technological University, Singapore;3. School of Computer Science and Engineering, Nanyang Technological University, Singapore;1. Departamento de Automática y Computación, Universidad Pública de Navarra, Campus Arrosadia s/n, 31006, Pamplona, Spain;2. Departamento de Matemáticas, Universidad Pública de Navarra, Campus Arrosadia s/n, 31006, Pamplona, Spain;3. Institute of Smart Cities, Universidad Pública de Navarra, 31006, Pamplona, Spain;1. Department of Computing Science and Mathematics, University of Stirling, UK;2. School of Computer Science and Engineering, Nanyang Technological University, Singapore;3. Computational Neuroscience and Functional Neurosurgery, University of Oxford, UK;2. Massachusetts Institute of Technology, Lexington, MA, United States
Abstract:Affective computing is an emerging interdisciplinary research field bringing together researchers and practitioners from various fields, ranging from artificial intelligence, natural language processing, to cognitive and social sciences. With the proliferation of videos posted online (e.g., on YouTube, Facebook, Twitter) for product reviews, movie reviews, political views, and more, affective computing research has increasingly evolved from conventional unimodal analysis to more complex forms of multimodal analysis. This is the primary motivation behind our first of its kind, comprehensive literature review of the diverse field of affective computing. Furthermore, existing literature surveys lack a detailed discussion of state of the art in multimodal affect analysis frameworks, which this review aims to address. Multimodality is defined by the presence of more than one modality or channel, e.g., visual, audio, text, gestures, and eye gage. In this paper, we focus mainly on the use of audio, visual and text information for multimodal affect analysis, since around 90% of the relevant literature appears to cover these three modalities. Following an overview of different techniques for unimodal affect analysis, we outline existing methods for fusing information from different modalities. As part of this review, we carry out an extensive study of different categories of state-of-the-art fusion techniques, followed by a critical analysis of potential performance improvements with multimodal analysis compared to unimodal analysis. A comprehensive overview of these two complementary fields aims to form the building blocks for readers, to better understand this challenging and exciting research field.
Keywords:
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号