首页 | 本学科首页   官方微博 | 高级检索  
     


Expressive Speech Animation Synthesis with Phoneme‐Level Controls
Authors:Z Deng  U Neumann
Affiliation:1. Computer Graphics and Interactive Media Lab, Department of Computer Science University of Houston, Houston, TX, USA zdeng@cs.uh.edu;2. Department of Computer Science, University of Southern California, Los Angeles, CA, USA uneumann@graphics.usc.edu
Abstract:This paper presents a novel data‐driven expressive speech animation synthesis system with phoneme‐level controls. This system is based on a pre‐recorded facial motion capture database, where an actress was directed to recite a pre‐designed corpus with four facial expressions (neutral, happiness, anger and sadness). Given new phoneme‐aligned expressive speech and its emotion modifiers as inputs, a constrained dynamic programming algorithm is used to search for best‐matched captured motion clips from the processed facial motion database by minimizing a cost function. Users optionally specify ‘hard constraints’ (motion‐node constraints for expressing phoneme utterances) and ‘soft constraints’ (emotion modifiers) to guide this search process. We also introduce a phoneme–Isomap interface for visualizing and interacting phoneme clusters that are typically composed of thousands of facial motion capture frames. On top of this novel visualization interface, users can conveniently remove contaminated motion subsequences from a large facial motion dataset. Facial animation synthesis experiments and objective comparisons between synthesized facial motion and captured motion showed that this system is effective for producing realistic expressive speech animations.
Keywords:Facial animation  speech animation  data‐driven  facial expression  phoneme‐isomap  motion capture  I  6  8 [Computing Methodologies]: Simulation and modelling –  types of simulation  I  3  7 [Computing Methodologies]: Computer graphics –  three dimensional graphics and realism
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号