Ma, XiaohanLe, Binh HuyDeng, ZhigangEitan Grinspun and Jessica Hodgins2016-02-182016-02-182009978-1-60558-610-61727-5288https://doi.org/10.1145/1599470.1599486Most of current facial animation editing techniques are frame-based approaches (i.e., manually edit one keyframe every several frames), which is ineffective, time-consuming, and prone to editing inconsistency. In this paper, we present a novel facial editing style learning framework that is able to learn a constraint-based Gaussian Process model from a small number of facial-editing pairs, and then it can be effectively applied to automate the editing of the remaining facial animation frames or transfer editing styles between different animation sequences. Comparing with the state of the art, multiresolution-based mesh sequence editing technique, our approach is more flexible, powerful, and adaptive. Our approach can dramatically reduce the manual efforts required by most of current facial animation editing approaches.Computer Graphics [I.3.7]Three Dimensional Graphics and RealismAnimationArtificial Intelligence [I.2.6]LearningAnalogiesStyle Learning and Transferring for Facial Animation Editing10.1145/1599470.1599486123-132