Automatic landmark annotation and dense correspondence registration for 3D human facial images
© Guo et al.; licensee BioMed Central Ltd. 2013
Received: 17 January 2013
Accepted: 15 July 2013
Published: 22 July 2013
Traditional anthropometric studies of human face rely on manual measurements of simple features, which are labor intensive and lack of full comprehensive inference. Dense surface registration of three-dimensional (3D) human facial images holds great potential for high throughput quantitative analyses of complex facial traits. However there is a lack of automatic high density registration method for 3D faical images. Furthermore, current approaches of landmark recognition require further improvement in accuracy to support anthropometric applications.
Here we describe a novel non-rigid registration method for fully automatic 3D facial image mapping. This method comprises two steps: first, seventeen facial landmarks are automatically annotated, mainly via PCA-based feature recognition following 3D-to-2D data transformation. Second, an efficient thin-plate spline (TPS) protocol is used to establish the dense anatomical correspondence between facial images, under the guidance of the predefined landmarks. We demonstrate that this method is highly accurate in landmark recognition, with an average RMS error of ~1.7 mm. The registration process is highly robust, even for different ethnicities.
This method supports fully automatic registration of dense 3D facial images, with 17 landmarks annotated at greatly improved accuracy. A stand-alone software has been implemented to assist high-throughput high-content anthropometric analysis.
Keywords3D face Facial morphology Registration Landmark localization Dense correspondence
Large-scale, high-throughput phenotyping is becoming increasingly important in the post-genomics era. Advanced image processing technologies are used more and more for collecting deep and comprehensive morphological data from different organisms, such as yeast , plants , worm  as well as mice ; and for different body parts such as brain [5, 6], lung  and face [8-13]. Especially for the brain 3D image registration, a novel elastic registration (HAMMER) of magnetic resonance images of the brain has greatly facilitated the medical research of brain [6, 14]. A recent work that combined florescent labeling and non-rigid registration achieved registration accuracy up to 2 μm in drosophila brain, which potentially allows functional analyses at individual neuron resolution . The soft tissue of the human face is a complex geometric surface composed of many important organs, including eyes, nose, ears, mouth, etc. Given its essential biological functions, the human face has been a key research subject in a wide range of fields including anthropology , medical genetics [8, 9, 16, 17], forensics [18, 19], psychology [20, 21], aging [22, 23] and the upcoming quantitative genomics [24, 25], etc.. Nonetheless, for a long period of time period the rich quantitative traits of face have not been made full use of. Previous anthropometric studies have been largely based on tedious manual measurements of dozens of distances between a set of landmarks, which were subjectively determined by the observers’ eyes and were thus error prone and sensitive to individual differences [26-28]. In the past few years, efforts have been paid to discover the genetic determinants of normal facial variations either by examining candidate genes or via genome-wide association studies [24, 25, 29-33]. Although high resolution 3D images were taken in some of these studies, the landmarks were still manually annotated ; and simple landmark coordinates or landmark-distances were used as the major phenotype data [24, 25, 33]. Such practices unavoidably led to loss of resolution and statistic power. In short, the lack of quantitative methods to capture the facial morphology in high definition and full automation has hindered the biological researches in human face. In fact, in the field of computer vision, many technologies have been developed for landmark recognition and dense point registration. Nonetheless, few have been successfully applied in the biological studies of human face. This is largely due to the different requirements between computer vision and biological studies. Quantitative biological analyses of face require the face registration to follow the principle of anatomical correspondence and landmarks have to be localized at high accuracies. However such rules are often not satisfied in the computer vision methods. For image registration, many methods rely on rigid transformation, such as the Iterative Closest Point (ICP) . ICP uses affine transformations, including rotation and translation to find the closest corresponding points between two surfaces. The registration based on ICP does not fully capture the anatomical variability, especially when faces to be compared differ significantly in shape or expression. For landmark localization, there exist many automatic methods [18, 35-39]. Landmark localization methods based on ICP suffer from the intrinsic incompatibility with anatomical correspondence [40, 41]. At some point, many landmark localization approaches use local, curvature-based facial features due to their invariance to surface translation and rotation. The two most frequently adopted features are the HK curvature and the shape index [35-37, 42]. However, curvature-based descriptors often suffer from surface irregularities, especially near eye and mouth corners . Other studies have used appearance-based methods where the facial features are modeled by basis vectors calculated from transformations such as Principal Component Analysis (PCA) [44, 45], Gabor wavelets [46, 47], or the Discrete Cosine Transform (DCT) . However, the lowest mean localization errors (root mean square error, RMS) these approaches can achieve were around 3 ~ 5 mm [18, 35-39], not accurate enough for high resolution morphometric analyses.
For biological inference, anatomical correspondence has to be established. This can be achieved by non-rigid transformations. A common method for deforming 3D surfaces is the thin-plate spline (TPS) algorithm . The process of using TPS warping involves minimizing a bending energy function for a transformation over a set of fiducial points (landmarks), thereby bringing the corresponding fiducial points on each surface into alignment with each other. A dense registration method has been developed based on TPS, and was successfully used to detect many face dysmorphology caused by rare genetic defects such as Noonan, 22q11 deletion, Bardet-Biedl and Smith-Magenis syndromes [8-13]. This approach therefore demonstrated great importance in biological and medical research. However, the TPS based registration has a key limitation that restrained its wide use in large-scale 3D facial datasets: A set of landmarks have to be localized as the anchoring points before TPS can be carried out. Methods have been developed to combine ICP-based landmark annotation and TPS warping to fully automate the registration [40, 41]. However, the landmark correspondences found by ICP are not exactly anatomically homologous, as previously discussed.
In this study, we develop an automatic registration method which combines a novel solution of landmark localization and an efficient protocol of TPS-based surface registration. The landmark localization mainly employs PCA to extract landmarks on surfaces by use of both shape and texture information. For the surface registration, a new TPS warping protocol that avoids the complication of inverse TPS warping (a compulsory procedure in the conventional registration method) is used to resample the meshes according to the reference mesh. We show that this method is highly accurate and robust accross different ethnicities. We also propose a new spherical resampling algorithm for re-meshing surfaces which efficiently removes the caveats and improves the mesh structure. Furthermore, the associated texture is also included in the registered data for visualization and various analyses.
Sample collection in this study was carried out in accordance with the ethical standards of the ethics committee of the Shanghai Institutes for Biological Sciences (SIBS) and the Declaration of Helsinki, and has been specifically surveyed and approved by SIBS. A written statement of informed consent was obtained from every participant, with his/her authorizing signature. The participants, whose transformed facial images are used in this study as necessary illustrations of our methodology, have been shown the manuscript and corresponding figures. Aside from the informed consent for data sampling, a consent of publication was shown and explained to each participant and their authorizing signature was obtained as well.
The 3D face data set
Preliminary nose tip localization and pose normalization
In 3D facial image processing, pose normalization and landmark localization are highly dependent on each other since pose normalization is typically guided by landmarks. The features commonly used for pose correction are the nose tip and inner eye corners as they are easier to detect , less sensitive to pose variation, and invariant to facial expressions [42, 44, 49, 50]. On the other hand, most existing landmark localization approaches rely on the assumption of frontal or approximately frontal poses and are therefore sensitive to roll and yaw rotation [38, 42, 51]. In order to fully automate the pose normalization and landmark annotation, we first identify the most robust and prominent landmark, the nose tip.
The pose correction becomes easy once the nose tip has been located. Correcting the pose basically consists of resetting the viewing coordinate system where an origin point and two axes must be defined. In some studies, the ICP matches are applied [52, 53]. Other studies try to find landmarks (i.e. inner eye corners) other than the nose tip to determine the pose [38, 48]. However, in this study we followed a rather practical solution in which all vertices within 50mm of the nose tip are used to correct the pose via the Hotelling transformation [52, 53].
Localization of the six most salient landmarks using PCA
P is then calculated for the signature eigenspace U using PCA (see Additional file 1 for details). To find the landmarks in a sample face, a patch of s mm×s mm is similarly defined for every point in the corresponding 2D grid, and a sample patch vector P s is derived following equation (2). P s is subsequently projected to the space U to evaluate its closeness to the origin point of U. In this study, two measurements of closeness are used, the reconstruction error e and the Mahalanobis distance d (see Additional file 1 for details). Sample points with smaller values for e and d are more likely to be a valid landmark. Therefore, the sample point corresponding to the minimum product value of e and d is defined as the landmark in our work. The patch size parameter s inevitably affects the final localization accuracy. We formally evaluated the dependence of the accuracy on the patch size, as illustrated in Additional file 3: Figure S2. Briefly, we checked the distances of the automatically annotated landmarks to the manually annotated ones, which we define as the localization error, for a random set of 100 individuals of different ethnicities and genders (25 individuals from each of the four groups: Han male, Han female, Uyghur male and Uyghur female). We found that in general, the localization error decreases with patch size (Additional file 3: Figure S2). However the error reaches a minimum when s is around 21mm and further increasing in the patch size does not reduce the error. Therefore we use the s value of 21mm throughout this study. To further optimize the computational efficiency, we narrow down the search for each landmark to a corresponding “landmark zone” on each sample face. Briefly, an arithmetic mean is calculated for each landmark across the training set, and projected onto the 2D space. Rectangular areas around the projection points are then defined as the landmark zones, with their sizes set experimentally (i.e. by training through a large number of faces) to ensure all real landmarks are encompassed. Therefore, the search for a particular landmark is done only within this landmark zone.
Heuristic localization of ten additional landmarks
Given the annotation of the six most salient landmarks, the pose of the surface can be fine tuned again. The reference plane is set to be the best fit plane to the six landmarks by least squares. The normal to the reference plane is set to be the z axis, and the y axis is given by the projection of the line going through the centers of lip corners and the eye corners onto the reference plane. The x axis is uniquely determined afterwards.
After the pose correction, 10 additional landmarks are identified heuristically by using geometric relations and texture constraints and the nose tip position is also updated. These 1andmarks include soft tissue nasion, alares, subnasale, labiale superius (upper lip point), stomion (the middle point between the upper and lower lip), labiale inferius (lower lip point), pogonion (chin point), and earlobe tips. The nose tip can be fine tuned according to the more uniformly defined coordinate system across all sample surfaces. Briefly, a semi-sphere is refitted around the previous nose tip and the point that minimizes the z coordinate error is chosen as the new nose tip. The subnasale point can be located by finding the inflection point with the minimum angle right below the nose tip. The alare points are the inflection points with the minimum local angles going horizontally away from the nose tip. Similar angle heuristics are applied to the detection of labiale superius, inferius, and stomion, with additional texture information in the YCbCr color space. For example, the labiale superius should locate the position on the border line where the Cr values below the line are greater (more red) than those above. Noticing that the region around the nasion point is approximately saddle-shaped and that of the chin point is ellipsoidal or sphere-shaped, both characterized by the two-way symmetry, we therefore locate the two points by finding the maximum local symmetry scores. The earlobe points are easily found by locating the tips with sheer slopes along the z-axis.
Spherical resampling and surface remeshing
The x-coordinate is multiplied by a factor before the coordinate conversion, to compensate for the face aspect ratio (height to width) . When plotted against θ and φ, the parameterized surface unfold into a nearly flat plane. This surface is then trimmed with an oval path to remove the irregular edges and re-sampled from a uniform square grid with an interval of 0.005 for both θ and φ. The re-sampled data points are then converted back to the Cartesian coordinate system to define a new surface mesh.
Surface registration for dense correspondence
In order to preserve the anatomical correspondence across the facial surfaces, we adopted the idea of the TPS-based registration method proposed previously . In that study, all surfaces were first manually annotated for a set of landmarks. The sample surfaces and the reference were all TPS warped to the cross-sample average landmarks. Each sample surface was then re-meshed by the closest points to the reference vertices, and further inverse TPS warped back to the original shape. Mathematically, TPS warping is not invertible. Although an approximation exists, it is computationally intensive and error prone . In our study, we designed an alternative scheme. First, a well-structured surface with few defects is chosen as the reference face, and spherically remeshed as described above. Then only the reference surface is TPS warped to each sample surface, taking the 17 landmarks as the fiducial points. The TPS warping is done as previously described . Thereafter the vertices on the reference surface find their closest projections on the sample surface, which define the new mesh vertices of the sample surface [13, 61]. The dense correspondence is established after all the sample surfaces are remeshed using the same reference. This approach eliminates the need for inverse TPS warping, and enhances the computational efficiency as well.
Accuracy of the landmark localization
In this section we demonstrate the accuracy of the proposed algorithm for automatic landmark localization. The accuracy is measured by the deviation of the automatically annotated landmarks from those manually annotated.
Mean error and standard deviation of 17 automatically annotated landmarks
Mean error (mm)
SD error (mm)
Mean error (mm)
SD error (mm)
Right eye outer corner
Right eye inner corner
Left eye inner corner
Left eye outer corner
Right lip corner
Left lip corner
Right earlobe tip
Left earlobe tip
Robustness of the registration method
The average faces calculation with the 3D face registration
In this work we propose a fully automatic registration method for high resolution 3D facial images. This method combines automatic landmark annotation and TPS-based registration. Pevious TPS based automatic registration methods suffer from low accuracy in landmark localization [40, 41], and are not suitable for anthropometric studies. For the problem of landmark localization, most time-honored solutions deal with only 2.5D data, leaving out the texture information. In particular, Perakis et al. described a method that made use of a comprehensive list of local shape descriptors, and achieved a precision of around 4 mm . Szeptycki et al. combined curvature analysis with a generic face model in a coarse-to-fine workflow, which enabled rotation invariant 3D landmark annotation at a precision of around 10 mm . On the other hand, D’Hose et al. made use of the Gabor wavelets to extract curvature information for coarse landmark localization, followed by an ICP-based fine mapping . This study achieved an overall precision level of a bit over 3 mm . Hutton et al. developed a method called the dense surface model (DSM), which hybridized the ICP optimization and active shape model (ASM) fitting to enable the automatic registration of 3D facial surfaces . They demonstrated that for the ten studied landmarks, the estimated positions using the DSM method have relatively small RMS errors (~3 mm) from the manual annotations. In this study, we constructed a novel PCA based landmark localization method, which made used of both the 3D geometric and 2D texture information, and achieved much lower landmark RMS errors, 1.7 ~ 1.8 mm on average, for a bigger number (17) of landmarks (Table 1). If the less salient landmarks, such as the earlobe tips, are excluded from the analysis, the errors will decrease further (Table 1). The novel use of both shape and texture information played a key role in improving the landmark localization accuracy. We found that the positions of some salient landmarks such as the eye corners are ambiguous even manually when the texture is removed. Texture gives rich information about the facial anatomical layout, such as the boundaries of different skin/tissue types. In fact, texture is almost the only information source for pattern recognition in 2D images and has been shown to give good performance. We projected both the shape and texture data into the 2D space, where the well-established PCA algorithm was used to detect the key landmarks. We also made use of the texture information for detecting certain other landmarks. Furthermore, due to the use of simple and optimized algorithms, the landmark annotation is also very efficient and does not require large amounts of memory. Hundreds of surfaces can be annotated within several minutes on a standard Windows PC. It is known that PCA can give wrong results on multi-modal feature distributions. This is particular the case when the surface data is taken under very different poses or light conditions, or image data contains non-Gaussian noises like spikes, holes and nevus. Under such conditions, PCA gives no guarantee that the modes of maximum variance (the principal component axis) are accounted for by the features of interest rather than noises. However, such problems are minimized in our study. First of all, for the purpose of anthropometric analyses, all surface images are supposed to be taken in laboratory environment where pose and light condition are well controlled to suppress unwanted variations. Moreover, several pre-processing and classification method were adopted to increase the specificity: 1. Median filter is used to reduce non-Gaussian noises; 2, both local patch vectors are normalized to have zero mean and unit length to eliminate bias to either modality; 3, the reconstruction error e was used along with the Mahalanobis distance for pattern classification. After all, the landmark recognition errors presented in Table 1 were assessed across all the individuals in the test panels. Unless obvious imaging errors were observed and the corresponding images removed, the errant landmark localizations were not specifically filtered from the test panels. The low average recognition errors therefore support the robustness of our method. The landmark localization may be further improved in the performance. For example, in the PCA based localization, one may apply bigger patch sizes but use a subset of points within each patch to construct the signature vector P. This may reduce the redundant information and result in better accuracy or efficiency. On the other hand, the gain in accuracy may also be partially attributed to the higher image resolution of our data (~30,000 vertices per surface on average) compared to the previous work (~10,000 vertices per surface). Furthermore, we also proposed a new protocol for the TPS-based registration, whereby the TPS warping was only applied to the reference face while the sample faces remained undeformed and thus avoided the step of inverse TPS warping, thereby further increasing the efficiency of our method. It is interesting to note that both the automatic landmark annotation and the TPS based registration steps work equally well for two different ethnicities, namely Han Chinese and Uyghur, in spite of the fact that they are substantially different in both genetic background and facial appearance. Han Chinese are representative of East Asian populations while Uyghur is an ancient admixture population whose ancestries came from both East Asians and Caucasians (European people) . As a result, Uyghur participants exhibited many Caucasian facial features such as sunken eyes and high nose ridge, etc. (Figure 6). This method was also tested on other ethnic groups and showed consistent robustness (data not shown). Such ethnicity independency is very important when this method is used to study the cross population facial morphological variations in humans.
It should be noted that the aim of this study is not to propose a general scheme of 3D surface registration. Rather, our method combines various pragmatic solutions to construct an optimized pipeline for high-throughput registration of dense 3D facial images. To the authors’ knowledge, this is the only fully automatic non-rigid registration method that aligns dense 3D face images at a landmark accuracy of ~ 2 mm. In fact, high resolution non-rigid registration methods that maximize the anatomical correspondence can greatly promote the biological and medical researches of the corresponding organs/features. A formal TPS based 3D face registration method, DSM, has revealed great potential in series of medical genetic studies of face dysmorphology [8-13]. In the field of brain research, efficient non-rigid methods, such as HAMMER , TPS-HAMMER  and Brainaligner  were developed for specific image data, and successfully applied in detection of aging/disease induced brain morphological changes  and delineation of ultrastructures  and neuronal circuits  of brain. Given the high definition and accuracy, our method may have many potential applications, such as quantitative characterization of human face diversity and detection of genetic/environmental factors that can induce facial shape changes. Furthermore, as this method is robust to diverse ethnicities, it is particularly suitable for studying evolution and divergence of human face among different populations. A standalone software has been implemented for this method, and is freely available for academic use upon request.
In the future, the anatomic correspondence can be further improved by including additional features such as the eyebrows, eyelid lines, and lip lines as landmarks. These features may provide discrimination power towards different facial expressions.
In summary, this study proposes a new scheme to build accurate and robust anatomical correspondence across dense surfaces of 3D facial images; and it was implemented into a fully automatic and efficient registration package. This method enables high-throughput capture and analysis of the wide ranging and yet fine detailed variations within human facial morphology. Such comprehensive and high resolution phenotypic data should be valuable in anthropological, disease diagnosis, and forensic studies of human facial morphology.
This work was supported by the Cutting Edge Research Project (Grant No. 2011KIP201) from the CAS-SIBS Outstanding Junior Researchers, the Key Research Direction Grant (No. KSCX2-EW-Q-1-12) from the CAS Knowledge Innovation Project and the Max-Planck-Gesellschaft Partner Group Grant. The funders had no role in study design, data collection or analysis, decision to publish, or preparation of the manuscript. Prof. Dr. Mark Stoneking from the Max Planck Institute for Evolutionary Anthropology has made valuable contributions to this study and provided proof-reading of the manuscript.
- Ohya Y, Sese J, Yukawa M, Sano F, Nakatani Y, Saito TL, Saka A, Fukuda T, Ishihara S, Oka S: High-dimensional and large-scale phenotyping of yeast mutants. Proc Natl Acad Sci USA. 2005, 102 (52): 19015-10.1073/pnas.0509436102.PubMed CentralView ArticlePubMedGoogle Scholar
- Tobias C, Roberto H, Falk S: HTPheno: an image analysis pipeline for high-throughput plant phenotyping. BMC Bioinforma. 2011, 12 (1): 148-10.1186/1471-2105-12-148.View ArticleGoogle Scholar
- Long F, Peng H, Liu X, Kim SK, Myers E: A 3D digital atlas of C. elegans and its application to single-cell analyses. Nat Methods. 2009, 6 (9): 667-672. 10.1038/nmeth.1366.PubMed CentralView ArticlePubMedGoogle Scholar
- Kristensen E, Parsons TE, Hallgramsson B, Boyd SK: A novel 3-D image-based morphological method for phenotypic analysis. Biomedical Engineering, IEEE Transactions. 2008, 55 (12): 2826-2831.View ArticleGoogle Scholar
- Peng H, Chung P, Long F, Qu L, Jenett A, Seeds AM, Myers EW, Simpson JH: BrainAligner: 3D registration atlases of Drosophila brains. Nat Methods. 2011, 8 (6): 493-500. 10.1038/nmeth.1602.PubMed CentralView ArticlePubMedGoogle Scholar
- Shen D, Davatzikos C: HAMMER: hierarchical attribute matching mechanism for elastic registration. IEEE Trans Med Imaging. 2002, 21 (11): 1421-1439. 10.1109/TMI.2002.803111.View ArticlePubMedGoogle Scholar
- Li B, Christensen GE, Hoffman EA, McLennan G, Reinhardt JM: Establishing a normative atlas of the human lung: intersubject warping and registration of volumetric CT images. Acad Radiol. 2003, 10 (3): 255-265. 10.1016/S1076-6332(03)80099-5.View ArticlePubMedGoogle Scholar
- Hammond P: The use of 3D face shape modelling in dysmorphology. Arch Dis Child. 2007, 92 (12): 1120-PubMed CentralView ArticleGoogle Scholar
- Hammond P, Hutton T, Allanson J, Buxton B, Karmiloff-Smith A, Patton M, Pober B, Smith A, Tassabehji M: 3D dense surface models identify the most discriminating facial features in dysmorphic syndromes. 2004, Toronto, Canada: 54th Annual Meeting of the American Society for Human GeneticsGoogle Scholar
- Hammond P, Hutton TJ, Allanson JE, Buxton B, Campbell LE, Clayton-Smith J, Donnai D, Karmiloff-Smith A, Metcalfe K, Murphy KC, et al: Discriminating power of localized three-dimensional facial morphology. Am J Hum Genet. 2005, 77 (6): 999-1010. 10.1086/498396.PubMed CentralView ArticlePubMedGoogle Scholar
- Hammond P, Hutton TJ, Allanson JE, Campbell LE, Hennekam RC, Holden S, Patton MA, Shaw A, Temple IK, Trotter M, et al: 3D analysis of facial morphology. Am J Med Genet A. 2004, 126A (4): 339-348. 10.1002/ajmg.a.20665.View ArticlePubMedGoogle Scholar
- Hutton TJ, Buxton BF, Hammond P: Automated registration of 3D faces using dense surface models. 2003, Citeseer: Proceedings of British Machine Vision Conference, 439-448.Google Scholar
- Hutton TJ, Buxton BF, Hammond P, Potts HWW: Estimating average growth trajectories in shape-space using kernel smoothing. Medical Imaging, IEEE Transactions on 2003. 2003, 22 (6): 747-753.View ArticleGoogle Scholar
- Klein A, Andersson J, Ardekani BA, Ashburner J, Avants B, Chiang MC, Christensen GE, Collins DL, Gee J, Hellier P, et al: Evaluation of 14 nonlinear deformation algorithms applied to human brain MRI registration. Neuro Image. 2009, 46 (3): 786-802.PubMed CentralPubMedGoogle Scholar
- Farkas LG, Katic MJ, Forrest CR: International anthropometric study of facial morphology in various ethnic groups/races. J Craniofac Surg. 2005, 16 (4): 615-10.1097/01.scs.0000171847.58031.9e.View ArticlePubMedGoogle Scholar
- Weinberg SM, Neiswanger K, Richtsmeier JT, Maher BS, Mooney MP, Siegel MI, Marazita ML: Three‒dimensional morphometric analysis of craniofacial shape in the unaffected relatives of individuals with nonsyndromic orofacial clefts: a possible marker for genetic susceptibility. Am J Med Genet A. 2008, 146 (4): 409-420.View ArticleGoogle Scholar
- Hammond P, Forster-Gibson C, Chudley AE, Allanson JE, Hutton TJ, Farrell SA, McKenzie J, Holden JJA, Lewis MES: Face and brain asymmetry in autism spectrum disorders. Mol Psychiatry. 2008, 13 (6): 614-623. 10.1038/mp.2008.18.View ArticlePubMedGoogle Scholar
- Albert AM, Ricanek K, Patterson E: A review of the literature on the aging adult skull and face: implications for forensic science research and applications. Forensic Sci Int. 2007, 172 (1): 1-9. 10.1016/j.forsciint.2007.03.015.View ArticlePubMedGoogle Scholar
- Kayser M, De Knijff P: Improving human forensics through advances in genetics, genomics and molecular biology. Nat Rev Genet. 2011, 12 (3): 179-192. 10.1038/nrg2952.View ArticlePubMedGoogle Scholar
- Meyer-Marcotty P, Alpers GW, Gerdes A, Stellzig-Eisenhauer A: Impact of facial asymmetry in visual perception: a 3-dimensional data analysis. Am J Orthod Dentofacial Orthop. 2010, 137 (2): 168-e161. 10.1016/j.ajodo.2009.09.005.View ArticlePubMedGoogle Scholar
- Little AC, Jones BC, DeBruine LM: Facial attractiveness: evolutionary based research. Philosophical Transactions of the Royal Society B: Biological Sciences. 2011, 366 (1571): 1638-1659. 10.1098/rstb.2010.0404.View ArticleGoogle Scholar
- Ramanathan N, Chellappa R, Biswas S: Computational methods for modeling facial aging: a survey. J Vis Lang Comput. 2009, 20 (3): 131-144. 10.1016/j.jvlc.2009.01.011.View ArticleGoogle Scholar
- Fu Y, Guo G, Huang TS: Age synthesis and estimation via faces: a survey. Pattern Analysis and Machine Intelligence, IEEE Transactions on 2010. 2010, 32 (11): 1955-1976.Google Scholar
- Liu F, van der Lijn F, Schurmann C, Zhu G, Chakravarty MM, Hysi PG, Wollstein A, Lao O, de Bruijne M, Ikram MA, et al: A genome-wide association study identifies five loci influencing facial morphology in Europeans. PLoS Genet. 2012, 8 (9): e1002932-10.1371/journal.pgen.1002932.PubMed CentralView ArticlePubMedGoogle Scholar
- Paternoster L, Zhurov AI, Toma AM, Kemp JP, St Pourcain B, Timpson NJ, McMahon G, McArdle W, Ring SM, Smith GD, et al: Genome-wide association study of three-dimensional facial morphology identifies a variant in PAX3 associated with nasion position. Am J Hum Genet. 2012, 90 (3): 478-485. 10.1016/j.ajhg.2011.12.021.PubMed CentralView ArticlePubMedGoogle Scholar
- Richardson ER: Racial differences in dimensional traits of the human face. Angle Orthod. 1980, 50 (4): 301-311.PubMedGoogle Scholar
- Allanson JE, O’Hara P, Farkas LG, Nair RC: Anthropometric craniofacial pattern profiles in down syndrome. Am J Med Genet. 1993, 47 (5): 748-752. 10.1002/ajmg.1320470530.View ArticlePubMedGoogle Scholar
- Dryden I, Mardia K: Statistical analysis of shape. 1998, Chichester: WileyGoogle Scholar
- Yamaguchi T, Maki K, Shibasaki Y: Growth hormone receptor gene variant and mandibular height in the normal Japanese population. Am J Orthod Dentofacial Orthop. 2001, 119 (6): 650-653. 10.1067/mod.2001.114536.View ArticlePubMedGoogle Scholar
- Tomoyasu Y, Yamaguchi T, Tajima A, Nakajima T, Inoue I, Maki K: Further evidence for an association between mandibular height and the growth hormone receptor gene in a Japanese population. Am J Orthod Dentofacial Orthop. 2009, 136 (4): 536-541. 10.1016/j.ajodo.2007.10.054.View ArticlePubMedGoogle Scholar
- Weinberg S, Naidoo S, Bardi K, Brandon C, Neiswanger K, Resick J, Martin R, Marazita M: Face shape of unaffected parents with cleft affected offspring: combining three‒dimensional surface imaging and geometric morphometrics. Orthod Craniofac Res. 2009, 12 (4): 271-281. 10.1111/j.1601-6343.2009.01462.x.PubMed CentralView ArticlePubMedGoogle Scholar
- Ermakov S, Rosenbaum MG, Malkin I, Livshits G: Family-based study of association between ENPP1 genetic variants and craniofacial morphology. Ann Hum Biol. 2010, 37 (6): 754-766. 10.3109/03014461003639231.View ArticlePubMedGoogle Scholar
- Boehringer S, Van Der Lijn F, Liu F, Günther M, Sinigerova S, Nowak S, Ludwig KU, Herberz R, Klein S, Hofman A: Genetic determination of human facial morphology: links between cleft-lips and normal variation. Eur J Hum Genet. 2011, 19 (11): 1192-1197. 10.1038/ejhg.2011.110.PubMed CentralView ArticlePubMedGoogle Scholar
- Besl PJ, McKay ND: A method for registration of 3-D shapes. IEEE Trans Pattern Anal Mach Intell. 1992, 14 (2): 239-256. 10.1109/34.121791.View ArticleGoogle Scholar
- Creusot C, Pears N, Austin J: Proceedings of the ACM workshop on 3D object retrieval. 3D face landmark labelling. 2010, New York, NY, USA: ACM, 27-32.Google Scholar
- Dibeklioglu H, Salah AA, Akarun L: 2nd IEEE International Conference on Biometrics: Theory, Applications and Systems: 2008. 3D facial landmarking under expression, pose, and occlusion variations. 2008, Arlington, VA, USA: IEEE, 1-6.Google Scholar
- Szeptycki P, Ardabilian M, Chen L: IEEE 3rd International Conference on Biometrics: Theory, Applications, and Systems: 2009. A coarse-to-fine curvature analysis-based rotation invariant 3D face landmarking. 2009, Washington, DC, USA: IEEE, 1-6.Google Scholar
- Chang KI, Bowyer W, Flynn PJ: Multiple nose region matching for 3D face recognition under varying facial expression. Pattern Analysis and Machine Intelligence, IEEE Transactions on 2006. 2006, 28 (10): 1695-1700.View ArticleGoogle Scholar
- Salah AA, Inar H, Akarun L, Sankur B: Robust facial landmarking for registration. Ann Telecommun. 2007, 62 (1-2): 1608-1633.Google Scholar
- Irfanoglu MO, Gokberk B, Akarun L: Proceedings of the 17th International Conference on Pattern Recognition: 2004. 3D shape-based face recognition using automatically registered facial surfaces. 2004, Washington, DC, USA: IEEE, 183-186.Google Scholar
- Schneider DC, Eisert P, Herder J, Magnor M, Grau O: Algorithms for automatic and robust registration of 3d head scans. Journal of Virtual Reality and Broadcasting. 2010, 7: 7-Google Scholar
- Colombo A, Cusano C, Schettini R: 3D face detection using curvature analysis. Pattern recognition. 2006, 39 (3): 444-455. 10.1016/j.patcog.2005.09.009.View ArticleGoogle Scholar
- Kakadiaris I, Passalis G, Toderici G, Murtuza N, Theoharis T: 3D face recognition. 2006, Edinburgh, UK: Proceedings of the British Machine Vision Conference: 2006, 869-868.Google Scholar
- Nair P, Cavallaro A: Matching 3D faces with partial data. 2008, Leeds, UK: Proc British Machine Vision Conference: 2008, 1-4.Google Scholar
- Abate AF, Nappi M, Riccio D, Sabatino G: 2D And 3D face recognition: a survey. Pattern Recognit Lett. 2007, 28 (14): 1885-1906. 10.1016/j.patrec.2006.12.018.View ArticleGoogle Scholar
- Wang Y, Chua CS, Ho YK: Facial feature detection and face recognition from 2D and 3D images. Pattern Recognit Lett. 2002, 23 (10): 1191-1202. 10.1016/S0167-8655(02)00066-1.View ArticleGoogle Scholar
- D’Hose J, Colineau J, Bichon C, Dorizzi B: First IEEE International Conference on Biometrics: Theory, Applications, and Systems. Precise localization of landmarks on 3d faces using gabor wavelets. 2007, Crystal City, VA, USA: IEEE, 1-6.Google Scholar
- Bookstein FL: Principal warps: thin-plate splines and the decomposition of deformations. Pattern Analysis and Machine Intelligence, IEEE Transactions on 1989. 1989, 11 (6): 567-585. 10.1109/34.24792. Washington, DC, USAView ArticleGoogle Scholar
- Sun Y, Yin L: 19th International Conference on Pattern Recognition: 2008. Automatic pose estimation of 3D facial models. 2008, Tampa, FL, USA: IEEE, 1-4.Google Scholar
- Lu X, Jain AK, Colbry D: Matching 2.5 D face scans to 3D models. Pattern Analysis and Machine Intelligence, IEEE Transactions on 2006. 2006, 28 (1): 31-43. Washington, DC, USAGoogle Scholar
- Salah AA, Alyz N, Akarun L: Registration of three-dimensional face scans with average face models. Journal of Electronic Imaging. 2008, 17: 011006-10.1117/1.2896291.View ArticleGoogle Scholar
- ter Haar FB, Veltkamp RC: A 3D face matching framework for facial curves. Graph Model. 2009, 71 (2): 77-91. 10.1016/j.gmod.2008.12.003.View ArticleGoogle Scholar
- Mian AS, Bennamoun M, Owens R: An efficient multimodal 2D-3D hybrid approach to automatic face recognition. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2007, 29 (11): 1927-1943.View ArticleGoogle Scholar
- Tsalakanidou F, Tzovaras D, Strintzis MG: Use of depth and colour eigenfaces for face recognition. Pattern Recognit Lett. 2003, 24 (9-10): 1427-1435.View ArticleGoogle Scholar
- Mian AS, Bennamoun M, Owens R: Keypoint detection and local feature matching for textured 3D face recognition. Int J Comput Vis. 2008, 79 (1): 1-12. 10.1007/s11263-007-0085-5.View ArticleGoogle Scholar
- Chang K, Bowyer K, Flynn P: Face recognition using 2D and 3D facial data. 2003, Santa Barbara, California, USA: ACM Workshop on Multimodal User Authentication: 2003, 25-32.Google Scholar
- Turk MA, Pentland AP: IEEE Computer Society Conference on Computer Vision and Pattern Recognition: 1991. Face recognition using eigenfaces. 1991, Maui, HI, USA: IEEE, 586-591.Google Scholar
- Alliez P, Ucelli G, Gotsman C, Attene M: Recent advances in remeshing of surfaces. Shape analysis and structuring. 2008, 53-82.View ArticleGoogle Scholar
- Hutton TJ, Buxton BR, Hammond P: IEEE Workshop on Mathematical Methods in Biomedical Image Analysis: 2001. Dense surface point distribution models of the human face. 2001, Kauai, HI, USA: IEEE, 153-160.Google Scholar
- Johnson H, Christensen G: Information Processing in Medical Imaging: 2001. Landmark and intensity-based, consistent thin-plate spline image registration. 2001, London, UK: Springer, 329-343.Google Scholar
- Schneider PJ, Eberly DH: Geometric tools for computer graphics. 2003, Morgan Kaufmann PubGoogle Scholar
- Gower JC: Generalized procrustes analysis. Psychometrika. 1975, 40 (1): 33-51. 10.1007/BF02291478.View ArticleGoogle Scholar
- Perakis P, Passalis G, Theoharis T, Kakadiaris IA: 3D Facial landmark detection & face registration. 2011, Tech Rep: University of AthensGoogle Scholar
- Xu S, Huang W, Qian J, Jin L: Analysis of genomic admixture in Uyghur and its implication in mapping strategy. Am J Hum Genet. 2008, 82 (4): 883-894. 10.1016/j.ajhg.2008.01.017.PubMed CentralView ArticlePubMedGoogle Scholar
- Wu G, Yap PT, Kim M, Shen D: TPS-HAMMER: improving HAMMER registration algorithm by soft correspondence matching and thin-plate splines based deformation interpolation. NeuroImage. 2010, 49 (3): 2225-2233. 10.1016/j.neuroimage.2009.10.065.PubMed CentralView ArticlePubMedGoogle Scholar
- Lao Z, Shen D, Xue Z, Karacali B, Resnick SM, Davatzikos C: Morphological classification of brains via high-dimensional shape transformations and machine learning methods. NeuroImage. 2004, 21 (1): 46-57. 10.1016/j.neuroimage.2003.09.027.View ArticlePubMedGoogle Scholar
- Yang JS, Awasaki T, Yu HH, He Y, Ding P, Kao JC, Lee T: Diverse neuronal lineages make stereotyped contributions to the Drosophila locomotor control center, the central complex. J Comp Neurol. 2013, 521 (12): Spc1-10.1002/cne.23366.Google Scholar
- Helmstaedter M, Mitra PP: Computational methods and challenges for large-scale circuit mapping. Curr Opin Neurobiol. 2012, 22 (1): 162-169. 10.1016/j.conb.2011.11.010.PubMed CentralView ArticlePubMedGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.