Joint facial expression recognition and intensity estimation based on weighted votes of image sequences

•A framework for joint facial expression recognition and intensity estimation from image sequences is proposed.•A feature representation based on weighted votes is also proposed.•Superior performance in estimating facial expression intensities.•Promising results with low computational complexity. Fa...

Full description

Saved in:
Bibliographic Details
Published inPattern recognition letters Vol. 92; pp. 25 - 32
Main Authors Kamarol, Siti Khairuni Amalina, Jaward, Mohamed Hisham, Kälviäinen, Heikki, Parkkinen, Jussi, Parthiban, Rajendran
Format Journal Article
LanguageEnglish
Published Amsterdam Elsevier B.V 01.06.2017
Elsevier Science Ltd
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:•A framework for joint facial expression recognition and intensity estimation from image sequences is proposed.•A feature representation based on weighted votes is also proposed.•Superior performance in estimating facial expression intensities.•Promising results with low computational complexity. Facial behavior consists of dynamically changing properties of facial features as a result of muscle activation. Facial behavior analysis is a challenging problem due to complexity of emotions and variability of the facial expressions associated with the emotions. Most facial expression recognition systems attempt to recognize facial expressions without taking into account the intensity of the expressions. In this paper, a novel framework for facial expression recognition and intensity estimation with low computational complexity requirement is proposed. The algorithm constructs a representation of facial features based on a weighted voting scheme and employs Hidden Markov Models to classify an input video into one of the six basic expressions, namely anger, disgust, fear, happiness, sadness, and surprise. The temporal segments, neutral, onset, and apex, of an expression are then obtained by means of a change-point detector. Evaluations on subject-independent analysis was conducted using Cohn-Kanade dataset and Beihang University facial expression datasets. The proposed approach has demonstrated a superior performance in recognizing facial expressions and estimating expression intensities.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0167-8655
1872-7344
DOI:10.1016/j.patrec.2017.04.003