Abstract
Communication plays a central role in everday life. During an average conversation, information is exchanged in a variety of ways, including through facial motion. Here, we employ a custom, model-based image manipulation technique to selectively “freeze” portions of a face in video recordings in order to determine the areas that are sufficient for proper recognition of nine conversational expressions. The results show that most expressions rely primarily on a single facial area to convey meaning with different expressions using different areas. The results also show that the combination of rigid head, eye, eyebrow, and mouth motions is sufficient to produce expressions that are as easy to recognize as the original, unmanipulated recordings. Finally, the results show that the manipulation technique introduced few perceptible artifacts into the altered video sequences. This fusion of psychophysics and computer graphics techniques provides not only fundamental insights into human perception and cognition, but also yields the basis for a systematic description of what needs to move in order to produce realistic, recognizable conversational facial animations.
Original language | English |
---|---|
Pages (from-to) | 251-269 |
Number of pages | 19 |
Journal | ACM Transactions on Applied Perception |
Volume | 2 |
Issue number | 3 |
DOIs | |
Publication status | Published - 2005 |
Keywords
- Applied perception
- Experimentation
- animation
- computer graphics
- facial expressions
- human-computer interface
ASJC Scopus subject areas
- Theoretical Computer Science
- General Computer Science
- Experimental and Cognitive Psychology