z-logo
open-access-imgOpen Access
Adaptive face modelling for reconstructing 3D face shapes from single 2D images
Author(s) -
Maghari Ashraf,
Venkat Ibrahim,
Liao Iman Yi,
Belaton Bahari
Publication year - 2014
Publication title -
iet computer vision
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.38
H-Index - 37
eISSN - 1751-9640
pISSN - 1751-9632
DOI - 10.1049/iet-cvi.2013.0220
Subject(s) - artificial intelligence , face (sociological concept) , computer science , facial recognition system , pattern recognition (psychology) , principal component analysis , computer vision , set (abstract data type) , image (mathematics) , iterative reconstruction , training set , model selection , solid modeling , social science , sociology , programming language
Example‐based statistical face models using principle component analysis (PCA) have been widely deployed for three‐dimensional (3D) face reconstruction and face recognition. The two common factors that are generally concerned with such models are the size of the training dataset and the selection of different examples in the training set. The representational power (RP) of an example‐based model is its capability to depict a new 3D face for a given 2D face image. The RP of the model can be increased by correspondingly increasing the number of training samples. In this contribution, a novel approach is proposed to increase the RP of the 3D face reconstruction model by deforming a set of examples in the training dataset. A PCA‐based 3D face model is adapted for each new near frontal input face image to reconstruct the 3D face shape. Further an extended Tikhonov regularisation method has been employed to reconstruct 3D face shapes from a set of facial points. The results justify that the proposed adaptive PCA‐based model considerably improves the RP of the standard PCA‐based model and outperforms it with a 95% confidence level.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here