Perceptually Valid Facial Expressions for Character-Based Applications

Peer reviewed: 
Yes, item is peer reviewed.
Scholarly level: 
Final version published as: 

International Journal of Computer Games Technology Volume 2009 (2009), Article ID 462315, 13 pages

Date created: 

This paper addresses the problem of creating facial expression of mixed emotions in a perceptually valid way. The research has been done in the context of a “game-like” health and education applications aimed at studying social competency and facial expression awareness in autistic children as well as native language learning, but the results can be applied to many other applications such as games with need for dynamic facial expressions or tools for automating the creation of facial animations. Most existing methods for creating facial expressions of mixed emotions use operations like averaging to create the combined effect of two universal emotions. Such methods may be mathematically justifiable but are not necessarily valid from a perceptual point of view. The research reported here starts by user experiments aiming at understanding how people combine facial actions to express mixed emotions, and how the viewers perceive a set of facial actions in terms of underlying emotions. Using the results of these experiments and a three-dimensional emotion model, we associate facial actions to dimensions and regions in the emotion space, and create a facial expression based on the location of the mixed emotion in the three-dimensional space. We call these regionalized facial actions “facial expression units.”

Document type: 
You are free to copy, distribute and transmit this work under the following conditions: You must give attribution to the work (but not in any way that suggests that the author endorses you or your use of the work); You may not use this work for commercial purposes; You may not alter, transform, or build upon this work. Any further uses require the permission of the rights holder (or author if no rights holder is listed). These rights are based on the Creative Commons Attribution-NonCommercial-NoDerivatives License.