Elsevier

Pattern Recognition

Volume 36, Issue 1, January 2003, Pages 259-275
Pattern Recognition

Automatic facial expression analysis: a survey

https://doi.org/10.1016/S0031-3203(02)00052-3Get rights and content

Abstract

Over the last decade, automatic facial expression analysis has become an active research area that finds potential applications in areas such as more engaging human–computer interfaces, talking heads, image retrieval and human emotion analysis. Facial expressions reflect not only emotions, but other mental activities, social interaction and physiological signals. In this survey, we introduce the most prominent automatic facial expression analysis methods and systems presented in the literature. Facial motion and deformation extraction approaches as well as classification methods are discussed with respect to issues such as face normalization, facial expression dynamics and facial expression intensity, but also with regard to their robustness towards environmental changes.

Introduction

Facial expression analysis goes well back into the nineteenth century. Darwin [1] demonstrated already in 1872 the universality of facial expressions and their continuity in man and animals and claimed among other things, that there are specific inborn emotions, which originated in serviceable associated habits. In 1971, Ekman and Friesen [2] postulated six primary emotions that possess each a distinctive content together with a unique facial expression. These prototypic emotional displays are also referred to as so called basic emotions. They seem to be universal across human ethnicities and cultures and comprise happiness, sadness, fear, disgust, surprise and anger. In the past, facial expression analysis was primarily a research subject for psychologists, but already in 1978, Suwa et al. [3] presented a preliminary investigation on automatic facial expression analysis from an image sequence. In the nineties, automatic facial expression analysis research gained much inertia starting with the pioneering work of Mase and Pentland [4]. The reasons for this renewed interest in facial expressions are multiple, but mainly due to advancements accomplished in related research areas such as face detection, face tracking and face recognition as well as the recent availability of relatively cheap computational power. Various applications using automatic facial expression analysis can be envisaged in the near future, fostering further interest in doing research in different areas, including image understanding, psychological studies, facial nerve grading in medicine [5], face image compression and synthetic face animation [6], video-indexing, robotics as well as virtual reality. Facial expression recognition should not be confused with human emotion recognition as is often done in the computer vision community. While facial expression recognition deals with the classification of facial motion and facial feature deformation into abstract classes that are purely based on visual information, human emotions are a result of many different factors and their state might or might not be revealed through a number of channels such as emotional voice, pose, gestures, gaze direction and facial expressions. Furthermore, emotions are not the only source of facial expressions, see Fig. 1. In contrast to facial expression recognition, emotion recognition is an interpretation attempt and often demands understanding of a given situation, together with the availability of full contextual information.

Section snippets

Facial expression measurement

Facial expressions are generated by contractions of facial muscles, which results in temporally deformed facial features such as eye lids, eye brows, nose, lips and skin texture, often revealed by wrinkles and bulges. Typical changes of muscular activities are brief, lasting for a few seconds, but rarely more than 5s or less than 250ms. We would like to accurately measure facial expressions and therefore need a useful terminology for their description. Of importance is the location of facial

Automatic facial expression analysis

Automatic facial expression analysis is a complex task as physiognomies of faces vary from one individual to another quite considerably due to different age, ethnicity, gender, facial hair, cosmetic products and occluding objects such as glasses and hair. Furthermore, faces appear disparate because of pose and lighting changes. Variations such as these have to be addressed at different stages of an automatic facial expression analysis system, see Fig. 2. We have a closer look at the individual

Representative facial expression recognition systems

In this section, we have a closer look at a few representative facial expression analysis systems. First, we discuss deformation and motion-based feature extraction systems. Then we introduce hybrid facial expression analysis systems, which employ several image analysis methods that complete each other and thus allow for a better overall performance. Multi-modal frameworks on the other hand integrate other non-verbal communication channels for improved facial expression interpretation results.

Discussion

In this survey on automatic facial expression analysis, we have discussed automatic face analysis with regard to different motion and deformation-based extraction methods, model and image-based representation techniques as well as recognition and interpretation-based classification approaches. It is not possible to directly compare facial expression recognition results of face analysis systems found in the literature due to varying facial action labeling and different test beds that were used

Conclusion

Today, most facial expression analysis systems attempt to map facial expressions directly into basic emotional categories and are thus unable to handle facial actions caused by non-emotional mental and physiological activities. FACS may provide a solution to this dilemma, as it allows to classify facial actions prior to any interpretation attempts. So far, only marker-based systems are able to reliably code all FACS action unit activities and intensities [58]. More work has to be done in the

Summary

In recent years, facial expression analysis has become an active research area. Various approaches have been made towards robust facial expression recognition, applying different image acquisition, analysis and classification methods. Facial expression analysis is an inherently multi-disciplinary field and it is important to look at it from all domains involved in order to gain insight on how to build reliable automated facial expression analysis systems. This fact has often been neglected in

About the Author—BEAT FASEL graduated from the Swiss Federal Institute of Technology Lausanne (EPFL) with a diploma in Communication Systems. He currently works towards a Ph.D. degree at IDIAP in Martigny, Switzerland. His research interests include computer vision, pattern recognition and artificial intelligence.

References (87)

  • G. Schwartz et al.

    Facial expression and imagery in depression: an electromyographic study

    Psychosomatic Med.

    (1976)
  • P. Ekman

    Emotions in the Human Face

    (1982)
  • P. Ekman et al.

    Facial Action Coding System: A Technique for the Measurement of Facial Movement

    (1978)
  • W. Friesen, P. Ekman, Emotional facial action coding system, unpublished manual,...
  • C. Izard, The maximally descriminative facial movement coding system (MAX), Available from Instructional Resource...
  • C. Izard, L. Dougherty, E. Hembree, A system for indentifying affect expressions by holistic judgments, unpublished...
  • N. Tsapatsoulis, K. Karpouzis, G. Stamou, A fuzzy system for emotion classification based on the MPEG-4 facial...
  • M. Hoch, G. Fleischmann, B. Girod, Modeling and animation of facial expressions based on B-splines, Visual Comput....
  • W. Friesen, P. Ekman, Dictionary—interpretation of FACS scoring, unpublished manuscript,...
  • P. Ekman, E. Rosenberg, J. Hager, Facial action coding system affect interpretation database (FACSAID),...
  • P. Ekman

    Methods for measuring facial actions

  • D. Matsumoto

    Cultural similarities and differences in display rules

    Motivation Emotion

    (1990)
  • D. Matsumoto

    Ethnic differences in affect intensity, emotion judgments, display rules, and self-reported emotional expression

    Motivation Emotion

    (1993)
  • A. Lanitis et al.

    Automatic interpretation and coding of face images using flexible models

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1997)
  • H. Hong, H. Neven, C. Von der Malsburg, Online facial expression recognition based on personalized galleries,...
  • J. Steffens, E. Elagin, H. Neven, PersonSpotter—fast and robust system for human detection, tracking and recognition,...
  • I. Essa et al.

    Coding, analysis, interpretation and recognition of facial expressions

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1997)
  • A. Pentland, B. Moghaddam, T. Starner, View-based and modular eigenspaces for face recognition, IEEE Conference on...
  • H. Rowley et al.

    Neural network-based face detection

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1998)
  • W. Fellenz, J. Taylor, N. Tsapatsoulis, S. Kollias, Comparing template-based, feature-based and supervised...
  • P. Belhumeur et al.

    Eigenfaces vs. fisherfaces: recognition using class specific linear projection

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1997)
  • M. Black, D. Fleet, Y. Yacoob, A framework for modeling appearance change in image sequences, Sixth International...
  • M. Black et al.

    Recognizing facial expressions in image sequences using local parameterized models of image motion

    Internat. J. Comput. Vision

    (1997)
  • M. Dailey, G. Cottrell, PCA Gabor for expression recognition, Institution UCSD, Number CS-629,...
  • C. Lisetti, D. Rumelhart, Facial expression recognition using a neural network, Proceedings of the 11th International...
  • M. Bartlett, Face image analysis by unsupervised learning and redundancy reduction, Ph.D. Thesis, University of...
  • J. Lien, Automatic recognition of facial expression using hidden Markov models and estimation of expression intensity,...
  • C. Padgett, G. Cottrell, Representing face image for emotion classification, in: M. Mozer, M. Jordan, T. Petsche...
  • G.W. Cottrell, J. Metcalfe, EMPATH: face, gender and emotion recognition using holons, in: R. Lippman, J. Moody, D....
  • T. Cootes et al.

    Active appearance models

    IEEE PAMI

    (2001)
  • G. Edwards, T. Cootes, C. Taylor, Face recognition using active appearance models, Proceedings of the Fifth European...
  • H. Kobayashi, F. Hara, Facial interaction between animated 3D face robot and human beings, Proceedings of the...
  • Z. Zhang, M. Lyons, M. Schuster, S. Akamatsu, Comparison between geometry-based and Gabor-wavelets-based facial...
  • Cited by (0)

    About the Author—BEAT FASEL graduated from the Swiss Federal Institute of Technology Lausanne (EPFL) with a diploma in Communication Systems. He currently works towards a Ph.D. degree at IDIAP in Martigny, Switzerland. His research interests include computer vision, pattern recognition and artificial intelligence.

    About the Author—JUERGEN LUETTIN received a Ph.D. degree in Electronic and Electrical Engineering from the University of Sheffield, UK, in the area of visual speech and speaker recognition. He joined IDIAP in Martigny, Switzerland, in 1996 as a research assistant where he worked on multimodal biometrics. From 1997 to 2000, he was head of the computer vision group at IDIAP, where he initiated and lead several European Community and Swiss SNF projects in the area of biometrics, speech recognition, face analysis and document recognition. In 2000, he joined Ascom AG in Maegenwil, Switzerland as head of the technology area Pattern Recognition. Dr. Luettin has been a visiting researcher at the Center for Language and Speech Processing at the Johns Hopkins University, Baltimore, in 1997 (large vocabulary conversational speech recognition) and 2000 (audio–visual speech recognition). His research interests include speech recognition, computer vision, biometrics, and multimodal recognition.

    View full text