TY - JOUR
T1 - Dynamics analysis of facial expression changes for person identification
AU - Tanaka, Hidenori
AU - Saito, Hideo
PY - 2010
Y1 - 2010
N2 - We propose a new method for analyzing dynamics of facial expression changes to identify persons. Several methods have been proposed to identify persons using facial images. In most methods, variations in facial expressions are one trouble factor because an input face image does not always contain the same facial expression as training images. However, the dynamics of facial expression changes are one measure of personal characteristics. In the proposed method, facial feature points are extracted using Active Appearance Models (AAMs) in the first frame of each video. They are then tracked using the Lucas-Kanade (LK) based feature point tracking method. Next, the starting and ending frames of facial expression changes are extracted by differences in the facial feature points' position between two successive frames. Finally, a feature vector is obtained as the sequence of the 2D coordinate variations of facial feature points. In the identification phase, an input feature vector is classified by calculating the distance between the input vector and the training vectors using dynamic programming matching (DP matching). We show the effectiveness of the proposed method using facial expression videos of the Facial Expressions and Emotions Database from Technical University of Munich (FEEDTUM database).
AB - We propose a new method for analyzing dynamics of facial expression changes to identify persons. Several methods have been proposed to identify persons using facial images. In most methods, variations in facial expressions are one trouble factor because an input face image does not always contain the same facial expression as training images. However, the dynamics of facial expression changes are one measure of personal characteristics. In the proposed method, facial feature points are extracted using Active Appearance Models (AAMs) in the first frame of each video. They are then tracked using the Lucas-Kanade (LK) based feature point tracking method. Next, the starting and ending frames of facial expression changes are extracted by differences in the facial feature points' position between two successive frames. Finally, a feature vector is obtained as the sequence of the 2D coordinate variations of facial feature points. In the identification phase, an input feature vector is classified by calculating the distance between the input vector and the training vectors using dynamic programming matching (DP matching). We show the effectiveness of the proposed method using facial expression videos of the Facial Expressions and Emotions Database from Technical University of Munich (FEEDTUM database).
KW - Facial expression analysis
KW - Person identification
UR - http://www.scopus.com/inward/record.url?scp=78650337469&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=78650337469&partnerID=8YFLogxK
U2 - 10.1541/ieejeiss.130.2047
DO - 10.1541/ieejeiss.130.2047
M3 - Article
AN - SCOPUS:78650337469
SN - 0385-4221
VL - 130
SP - 2047-2057+21
JO - IEEJ Transactions on Electronics, Information and Systems
JF - IEEJ Transactions on Electronics, Information and Systems
IS - 11
ER -