Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

An automatic approach for classification and categorisation of lip morphological traits

Abstract

Classification of facial traits (e.g., lip shape) is an important area of medical research, for example, in determining associations between lip traits and genetic variants which may lead to a cleft lip. In clinical situations, classification of facial traits is usually performed subjectively directly on the individual or recorded later from a three-dimensional image, which is time consuming and prone to operator errors. The present study proposes, for the first time, an automatic approach for the classification and categorisation of lip area traits. Our approach uses novel three-dimensional geometric features based on surface curvatures measured along geodesic paths between anthropometric landmarks. Different combinations of geodesic features are analysed and compared. The effect of automatically identified categories on the face is visualised using a partial least squares method. The method was applied to the classification and categorisation of six lip shape traits (philtrum, Cupid’s bow, lip contours, lip-chin, and lower lip tone) in a large sample of 4747 faces of normal British Western European descents. The proposed method demonstrates correct automatic classification rate of up to 90%.

Introduction

The face is the most expressive part of the human body and is essential in everyday social interaction. The lips are one of the key components of the face (Fig 1); the lip area runs from the base of the nose to the tip of the chin, and therefore constitutes most of the lower third of the face. The lips include the philtrum and Cupid’s bow. The lip vermillion is the thin layer of skin, red in colour, overlying a highly vascularized region. The appearance of the lips varies with facial movement; therefore, for accurate anthropometric measurement the lips should be assessed when the subject is relaxed and has a natural head posture [1]. The morphological features of the lips vary greatly between individuals and are particularly dependent on age, sex and ethnicity [2, 3].

Studying lip morphology is important for many applications, including face recognition and gender classification [47]. The lips have also been reported to contribute to facial attractiveness and a number of studies have attempted to evaluate lip aesthetics by creating norms and standards of ideal lip position and shape based on cephalometric analysis and facial measurements [811]. Furthermore, lip morphology plays an important role in diagnosis and analysis of many medical conditions and facial dysmorphology. It is also important to find genetic variants associated with facial syndromes.

Cleft lip (CL) is one of the most recognisable facial anomalies which has been the focus of clinical research for many decades [12]. The association between genetic variants and environmental factors associated with cleft lip has also been reported [13]. Fetal alcohol syndrome (FAS) is another medical condition where lip morphology is used in a diagnostic role. FAS features are smooth philtrum, thin upper lip vermilion and short palpebral fissure length [14].

Thus, the identification and classification of lip shape characteristics is important in medical practice. So far, the categorisation (clustering) of lip morphological characteristics (traits) has mostly been performed subjectively in clinical practice, which is time consuming and prone to error. In this paper, we propose a method for automatic classification and categorisation of lip morphological traits. We investigate six shape traits of the lip area (Philtrum, Cupid’s bow, lip contours, chin, and lower lip tone) using the proposed approach (Fig 2). These traits were categorised subjectively by medical experts and enables a direct comparison with an automated approach.

thumbnail
Fig 2. Lip traits.

Description of the lip traits used in the present study.

https://doi.org/10.1371/journal.pone.0221197.g002

The proposed method relies on a combination of 3D geometric features, which exhibited robust results for gender classification in our previous study [15]. These features are derived from the mean and Gaussian local curvatures, shape indices, and curvedness measures obtained for geodesic path (shortest path) between anthropometric landmarks in the lip area.

The main contributions of this work can be summarized as follows:

  • It proposes an approach for automatic categorisation and classification of lip morphology. To the best of our knowledge, no such automatic approach has been suggested so far.
  • It employs novel geodesic curvature features for the description of lip shape.
  • It adopts a new method, based on partial least squares regression, for visualising the effect of the trait categories on the lip region.
  • It compares the results of the automatic categorisation with those of the subjective(manual) categorisation previously reported in [16]. The supervised classification rates based on the automatic categories outperform those based on the manual categories by at least 8%. In addition, the automatic labels were found to be more efficient than the manual labels.

Related work

The traditional methods that have been employed to assess and categorise lips morphological traits can be divided into three categories: clinical examination, ordinary two-dimensional (2D) photography, and three-dimensional (3D) facial imaging.

In clinical examination, certain features (e.g., lip width, height) are analysed for the subject by direct observation or measurement performed by a medical expert [17, 18]. This method is accurate to some degree but need specific instruments and trained expert clinician to make the measurements.

Ordinary 2D photography is the most common method of recording human’s facial morphology due to the lost cost, ease of use and availability of the necessary equipment. For example, in the past, 2D photographs of faces were used to obtain anthropometric measurements and classification of lip morphological traits [19, 20]. However, 2D photographs do not capture 3D face shape, which contains important morphological information.

In the last two decades, 3D imaging (e.g. 3D meshes obtained using laser or other imaging technologies) has become more common in various medical applications. 3D facial images are much more informative than 2D photographs [21, 22] and thus can be more useful for studying lip morphology. For example, Wildon et al. [23] used 3D facial images of 109 subjects, aged 5–6 years, to produce four categories for philtrum shape: triangular, parallel, concave, and flat. Later on, Wilson et al. [16] produced 3D measurements of lip vermilion and Cupid’s bow and described different morphological features of the vermilion of the lips and associated lip traits for 4747 subjects from the ALSPAC dataset [16]. In Addition, Lee et al. [24] used the 3D face to identify measures of facial appearance for designing reconstructive surgery for Cleft lip with or without cleft palate of Hispanic/Latino White children. Although past research indicates the popularity of 3D imaging for research in facial morphology, current methods in the area rely on solely manual facial trait classification and categorisation, which is a very time consuming process. Consequently, such research would benefit significantly from an automatic approach.

In the data mining world, automatic categorisation (clustering) and classification are two types of learning methods. Both these methods characterise objects into groups based on one or more features.The key difference between categorisation and classification is that categorisation is an unsupervised learning technique used to group similar instances on the basis of features whereas classification is a supervised learning technique used to assign predefined tags to instances on the basis of features [25]. Generally, the performance of the clustering or classification algorithm depends heavily on the nature of processed dataset [26, 27]. In this paper, Support Vector Machine (SVM) [28] and boosting algorthim [29] are used for automatic classification and Kmeans++ [30] for automatic categorisation.

The majority of previous studies in facial morphology used simple geometric features such as the Euclidean distances and angles between landmarks [31, 32]. Nevertheless, 3D facial meshes contain far richer information related to the 3D shape of the faces. Therefore, new informative surface shape features are employed in this paper for the classification and categorisation of lips morphological traits. These features are derived from the mean and Gaussian curvatures, shape indices, and curvedness measures calculated at certain points along the geodesic path between 3D facial anthropometric landmarks. The mean, max and min curvatures have been successfully utilised to classify philtrum morphology in the past [33]. The shape index and curvedness measures have been successfully applied in a variety of 3D face recognition applications [3437]. In addition, the geodesic paths have been widely used in face recognition (FR) systems for faces with different poses and expressions (e.g., see [3840]). The above studies employed the radial geodesic paths or iso-geodesic paths of the whole face as features for FR purposes. In the present study, the geodesic paths are identified between the anthropometric 3D face landmarks (see Section 3.3.1).

Many studies have used partial least squares (PLS) regression [41] for analysing the effects and determine the statistical significance of biological and environmental variables such as age, gender, ethic and BMI on face module [4244]. Consequently, we adopted PLSR with dummy variables for visualising the influence of the discovered categories on the facial physical appearance to gain insight into suitability of categories for description of the underlying facial traits.

Materials and methods

Dataset, landmarks and manual categorisation of lip traits

This study is based on three-dimensional facial data collected from 15-year-old children from the Avon Longitudinal Study of Parents and Children [45]. Ethical approval for the study was obtained from the ALSPAC Ethics and Law Committee and the Local Research Ethics Committees (UBHT): 06/Q2006/531 Avon Longitudinal Study of Parents and Children (ALSPAC), Hands on Assessments: Teen Focus 3 (Focus 15+) (7th August 2006;Confirmed 15th September 2006). Written consent was also obtained from parents and guardians prior to obtaining the facial scans. This prospective study recruited pregnant women living in the former county of Avon in South-West England with an estimated delivery date of between April 1st 1991 and December 31st 1992. The initial number of pregnancies enrolled was 14,541 (for these at least one questionnaire was returned or a “Children in Focus” clinic had been attended by 19/07/99). Of these initial pregnancies, there were a total of 14,676 fetuses, resulting in 14,062 live births and 13,988 children who were alive at 1 year of age. Please note that the study website contains details of all the data that is available through a fully searchable data dictionaryand variable search tool REF (http://www.bris.ac.uk/alspac/researchers/data-access/data-dictionary/) The children were invited to a research clinic when they were 15 years old. A subset of 4,747 children (2,233 males, 2,514 females) attended this clinic and had a three-dimensional facial scan taken using two Konica Minolta Vivid 900 laser cameras [46]. The reliability of image capture has been reported extensively elsewhere [47]. After that, 21 facial landmarks represented by a total of 63 (x, y, and z) coordinate values were manually identified (Manually obtained landmarks were already available for this dataset. However, they can be obtained automatically using one of the methods described in [4850].) and recorded for each 3D facial image [47]. The biological landmark points for the lip region on the human face are illustrated in Fig 3. Table 1 explains the definitions of the landmarks. S1 Table records the 21 facial landmarks (x, y, and z) values for all ALSPAC data set subjects.

thumbnail
Fig 3. Lip landmarks.

There are eight anthropometric landmarks localized in the lip region.

https://doi.org/10.1371/journal.pone.0221197.g003

thumbnail
Table 1. Biological definitions of soft tissue landmarks in the lip region.

https://doi.org/10.1371/journal.pone.0221197.t001

Data used for this research will be made available on request to the ALSPAC executive committee (alspac-exec@bristol.ac.uk). The ALSPAC data management plan (available here: https://proposals.epi.bristol.ac.uk) describes in detail the policy regarding data sharing, which is through a system of managed open access.

This paper deals with six lip traits previously identified in [16]. Specifically, these are philtrum shape, Cupid’s bow shape, upper and lower lip contours, lateral lip-chin shape, and lower lip tone shape (Fig 2). Each trait was manually categorised into three to seven categories depending on its appearance and the respective labels were assigned to all ALSPAC images [16].

Proposed algorithm

In this subsection, we describe the approach we use for automatic classification and categorisation of lip morphology. To the best of our knowledge, it represents the first automatic algorithm ever developed for these purposes, while previous approaches relied on manual work of highly trained clinicians. An automated approach has the benefit of bringing technology to a wider general use and opening up a potential for analysis of large datasets with direct advantages to research areas such as finding genetic associations and automatic diagnosis of facial syndromes.

Figs 4 and 5 display block diagrams of the proposed approach, which are explained below.

thumbnail
Fig 4. Classification block diagram.

This diagram illustrates the proposed approach for automatic classification lip traits.

https://doi.org/10.1371/journal.pone.0221197.g004

thumbnail
Fig 5. Categorisation block diagram.

This diagram illustrates the proposed approach for automatic categorisation lip traits.

https://doi.org/10.1371/journal.pone.0221197.g005

Initial processing of facial images.

The facial data captured by 3D scanners frequently suffers from imperfections. These imperfections are generally in the form of noise, unwanted elements (such as hair, eyelashes, ears, or portions of the neck or clothes), or holes due to missing data at dark, shiny, or hairy regions (such as eyebrows, beard, the iris, or open mouth). The removal or correction of the imperfections is an essential preliminary stage of data analysis, usually called preprocessing. In addition, the preprocessing often includes remeshing of the resulting images, to ensure that the mesh becomes more uniform and has no defects, and normalisation of the face posture, to ensure that all faces are in the same position [51]. For further analysis, a regularisation of the number and position of vertices can also be made through finding dense correspondences in the mesh [52, 53].

In this paper, two approaches for 3D face preprocessing are implemented for comparing their effects on lip area morphological traits classification, one without the regularisation and the other with the regularisation.

Preprocessing without mesh regularisation.

All raw images were processed using the Rapidform 2006 software following the steps mentioned above [54] except for regularisation of the mesh. Specifically, the steps taken were:

  • Denoising. The unwanted elements of the scans were removed manually followed by automatic removal of large spike, which can naturally arise in scanning. The images were further smoothed using the volume-preserving Laplacian filter available in Rapidform, which removes little spikes. Laplacian smoothing is a common procedure (e.g., see [5456]), which is implemented in many software packages including, for example, Graph MATLAB Toolbox [57].
  • Hole filling. Small or large holes may arise in scanning and denoising. All holes were automatically identified and filled using Rapidform built-in tools with curvature preserving option. we can also use 3DFaceModelsPreprocessingTool1 [58], to eliminate noise and remove undesired parts of the face and fill the holes automatically.
  • Registration and merging. All faces were obtained using two laser cameras generating two scans of the face, left and right [59]. After denoising and hole filling, the left and right images were registered together using the Rapidform best-fit tool followed by merging to create a single facial image. The mesh quality was further checked and defects removed automatically.
  • Normalisation. To insure that all faces are in the same position, all images were manually landmarked [47] and normalised by fitting them to a vertical cylinder [51], with mid-endocanthion used as the origin of coordinates.

Images obtained following these steps will be referred to as non-regularised.

Preprocessing with mesh regularisation.

In addition to the above steps, all non-regularised images were superimposed on a reference face with 7150 vertices and regularised to ensure that the resulting meshes have exactly 7150 vertices each. Specifically, the following was done:

  • The preprocessing steps without mesh regularisation are repeated.
  • Non-rigid registration. The idea was to use an anthropometric facial template with 7150 vertices to crop the facial images of interest. The template was an average face constructed from 3D facial images of 400 Western Australian healthy young individuals, aged 5–25 years, captured with a 3dMD imaging system [53]. The template was mapped onto the facial meshes using an iterative closest point (ICP) algorithm [60].
  • Regularisation. An iterative procedure was used to assign each floating point of the template to a corresponding point of the target image by applying the distance weighted k-nearest neighbor rule [61]. As a result, all faces were cropped and converted to uniformly distributed meshes each consisting of 7150 vertices.

The resulting images will be referred to as regularised.

It must be noted that the non-regularised images were of much higher resolution (50 to 70 thousand vertices each) than the regularised images (exactly 7150 vertices each). The regularised images were further used to test the robustness of our 3D geometric features as well as the classification and categorisation systems on low-resolution meshes in comparison with the non-regularised images.

Fig 6 illustrates the difference between an non-regularised and a regularised face. It is apparent that the lip morphology is more prominent in non-regularised faces. However, regularised images allow for faster extraction of facial features. Furthermore, the regularisation allows us to test our automatic classification method for both high-resolution images (e.g., captured with laser scanners) and low-resolution images (e.g., obtained using structured light stereo or photogrammetric scanners).

thumbnail
Fig 6. Difference between preprocessed images.

(A) Non-regularised face. (B) Regularised face.

https://doi.org/10.1371/journal.pone.0221197.g006

Feature extraction and normalisation.

Previous studies have shown advantages of using curvatures in 3D facial applications. This prompted us to use curvatures to classify and cluster facial morphological traits, too; however, we adopted a different, novel strategy. Specifically we decided to use geodesic paths between anthropometric landmarks to define key points at which curvatures can be calculated and combine these curvatures into feature descriptors for the anatomical facial traits in the lip region (Cupid’s bow shape, lip contours, etc.).

We also calculated the Euclidean and geodesic distances to assess their efficiency for the classification and categorisation of lip traits as compared to combinations of geodesic curvature features. This is because many previous studies frequently used these quantities as features for 3D facial morphology analysis (e.g., see [31, 62, 63]).

Extracting geodesic paths.

The geodesic path is the shortest route between two points on a surface and the geodesic distance is the length of this route [64]. There are a number of algorithms to compute geodesic paths and distances on triangular meshes; some are approximate, such as the fast marching method [65], while others are exact (however, relatively slow). The exact algorithms include, for example, the Mitchell–Mount–Papadimitriou (MMP) [66], Chen–Han (CH) [67] methods and parallel fast marching algorithm [68]. Exact methods are more accurate in tracking geodesic paths than approximate methods and are especially advantageous for low-resolution meshes. Fig 7 highlights the difference between an exact and a fast geodesic algorithm in determining geodesics in a synthetic low-resolution mesh, where the black (exact) trajectory clearly follows the mesh edges more regularly than the red (fast) trajectory.

thumbnail
Fig 7. Exact versus fast marching algorithm.

Extracting geodesic paths using an exact and a fast method for a synthetic mesh.

https://doi.org/10.1371/journal.pone.0221197.g007

Gabriel Peyre’s MATLAB fast-marching toolbox [69] and the exact geodesic toolbox [70] were used to find geodesic paths and calculate the geodesic distances between two landmarks. The former toolbox was used for high-resolution data (non-regularised meshes), while the latter one was used for low-resolution data (regularised meshes), because, as shown in Fig 7, the exact method is more suitable for tracking low-resolution meshes and this is compatible with the finding of [71].

Fig 8A illustrates the paths used for all lip traits (see Table 2), apart from the lower lip tone trait. For the lower lip tone, we used the geodesic path between the lower lip contour landmarks (chL, li and chR) and four extra geodesic paths connecting the three points obtained from the above landmarks by shifting them down by a certain distance (Fig 8B).

thumbnail
Fig 8. The geodesic paths used for the classification and categorisation of the lip traits.

(A) Paths for the lip traits. (B) Paths for the lower lip tone.

https://doi.org/10.1371/journal.pone.0221197.g008

thumbnail
Table 2. List of geodesic paths defining morphological lip traits.

https://doi.org/10.1371/journal.pone.0221197.t002

Curvature features.

The local principal curvatures were first calculated at the vertices of the geodesic path. The mean curvature, Gaussian curvature, shape index, and curvedness were then calculated from the principal curvatures.

The principal curvatures result from the intersection of the 3D surface with a orthogonal to the tangential plane. Many methods have developed to calculate the principal curvatures for the 3D images [72, 73]. In this work the Normal Cycle curvature tensor method is used to calculate the principal curvatures for the 3D images. This method was implemented, in particular, by [74] to give a general method to define principal curvatures for the 3D images. The principal curvatures k1 and k2 at v are estimated by the eigenvalues of normal cycle theory based equation , while the eigenvectors represent the curvature directions [75]. (1) where v represents the vertex position on the mesh, |B| is the surface area around v over which the curvature tensor is estimated, β(e) is the signed angle between the normal vector to the two oriented triangles incident to edge e, |eB| is the length of eB, and e is a unit vector in the same direction as e [75].

Using principal curvatures (k1 and k2), Mean (M), Gaussian (Ga) curvature, Shape index (Sh) and curvedness (C) are calculated as: (2) (3) (4) (5)

Shape index (Sh) quantitatively measures the shape of a surface at a vertex v and captures the intuitive notion of local shape of a surface. while, curvedness (C), measures how highly or softly bent a surface is. Curvedness can define the scale difference between objects: for example the difference between a soccer ball and a cricket ball. See [33] and [76] for more information on the principal curvature calculation algorithm.

Calculation of the geodesic and Euclidean distances.

For each lip trait, the geodesic distances and Euclidean distances were calculated between the same landmarks as those utilised to extract the geodesic paths shown in Fig 8a. The fast marching algorithm and exact algorithm were used to compute geodesic distances from high-resolution and low-resolution faces, respectively. The Euclidean distance between points A and B in three dimensions, defined by their coordinates (XA, YA, ZA) and (XB, YB, ZB), is calculated as (6)

Normalisation of the curvature features.

Histogram normalisation is used to normalize the features descriptors. It is an estimate of the probability distribution of a continuous variable. It is a kind of bar graph, to construct a histogram, the first step is to “bin” the range of values—that is, divide the entire range of values into a series of intervals—and then count how many values fall into each interval. The bins are usually specified as consecutive, non-overlapping intervals of a variable. If the bins are of equal size, a rectangle is erected over the bin with height proportional to the frequency—the number of cases in each bin. A histogram may also be normalized to display “relative” frequencies. It then shows the proportion of cases that fall into each of several categories, with the sum of the heights equaling one [77]. In this work, each geodesic path has a different number of nodes (vertices) for curvature calculation. To deal with this, a normalised histogram distribution was calculated for each path feature; for this purpose, the number of bins selected was 5, 10, 15, 20, or 25, depending on the minimum number of nodes in a path across the entire sample, the node numbers for the longest path (li-pg) were around 100 node. The histogram representation has used for features normalisation and representation in many image and pattern recognition such as Bag-of-Words (BoW) image representation [78], this motivate us to use the histogram representation to normalise the geodesic path curvature features.

The number of vertices defining a geodesic path varies between faces. A normalisation procedure is required to ensure that the respective paths of all facial images have the same number of points at which curvature features are measured. To this end, a histogram distribution was calculated for each feature; the number of bins selected was 5, 10, 15, 20, or 25, depending on the maximum and minimum number of points in the path.

Let denote the vertices of a path Pk on facial mesh k and let , , , and denote, respectively, the mean curvature, Gaussian curvature, curvedness value, and shape index value evaluated at vertex . For each path, we choose a number b = 5, 10, 15, 20, or 25 such that b ≤ min n, where min n is the minimum number of vertices in all paths Pk across the sample. After the histogram normalisation (using the MATLAB function histnorm) with b bins, we get exactly 4b characteristic curvature features for path Pk: (7) (8) (9) (10) where ^ denotes the respective values resulting from the histogram normalisation. Then a features descriptor is composed Dk = [Mk, Gak, Ck, Shk] consisting of 4b components. This procedure provided us with four vectors of equal length: mean curvature, Gaussian curvature, shape index, and curvedness. These vectors were then concatenated to produce a single vector, feature descriptor, for each path in a face; see Fig 8 and Table 2.

Data balancing and classification.

The problem of imbalanced datasets can arise in classification when the number of elements in one class is much lower than that in other classes. Standard classifiers tend to overestimate the importance of the larger classes and underestimate the importance of the smaller classes. To cope with this problem, several methods have been suggested [7981].

The present study uses the boosting method [29] to classify the unbalanced ALSPAC dataset. We also compare our results with those obtained using the multiclass SVM (supper vector machine) method [28], which does not involve data balancing. In this work, the public MATLAB Software called Multiclass Gentle Adaboosting was used for classification purposes [82].

Automatic categorisation.

The present study carries out automatic categorisation of lip traits by using the Kmeans++ clustering technique [30, 83]. Kmeans++ function on the same rule of kmeans for data clustering by choosing K initial centroids, where K is a user specified parameter. kmeans algorithm concepts is very simple, it is work on minimizing the sum of squared distances from a cluster center, to the cluster members.

The kmeans algorithm proceeds major steps are:

  • initial centroids K is selected randomly.
  • The group of points are assigned to nearest cluster according to Euclidean distance metric.
  • The centroids to the mean of the members of the cluster is updated.
  • The second and the third steps are repeated until the assignments from the second step do not change [30].

Kmeans++ handle the problem of choosing the cluster center randomly by performing a procedure to initialize the cluster centers before implementing the standard kmeans optimization iterations. The Kmeans++ initialization steps are:

  • First centroid is chosen randomly.
  • the distance for each data point to the nearest centroid that has already been chosen is computed.
  • the next centroid is selected using a weighted probability proportional to distance value.

Using the Kmeans++ initialization procedure provides more accurate cluster centroid with minimum time [83].

To evaluate the clustering performance and find an optimum number of clusters, we used three different internal validation indices: silhouette index (SI), Dunn index (DI), and Calinski–Harabasz index (CH) [84, 85]. Internal validity indices rely on properties intrinsic to the data set. Most index measures are based on the concept that the points in the same cluster should be similar and the points in different clusters should be dissimilar. Below we briefly discuss how these concepts are defined for each of the three internal index measures:

  • The silhouette index (SI) validates the clustering performance based on the pairwise difference of between-cluster and within-cluster distances. Moreover, the optimal cluster number is determined by maximizing the value of this index. This index, denoted s(i), is computed as (11) where a(i) is the average distance between the ith element and all other elements within the same cluster, while b(i) is the minimum average distance between the element ith and any other cluster, of which the ith element is not a member [84, 86].
  • The Calinski–Harabasz index (CH) evaluates the cluster validity based on the average between-cluster and within-cluster sum of squares. CH index calculates separation based on the maximum distance between cluster centroids, and measures compactness(Compactness measures how closely the objects in a cluster are related [86]) depending on the sum of distances between objects and their cluster center. In addition, the optimal cluster number is determined by maximizing the value of this index. This index is computed as (12) where SB is the between-cluster scatter matrix, SW is the internal (within-cluster) scatter matrix, np is the number of clustered samples, and k is the number of clusters [84, 85].
  • The Dunn index (DI) measures the minimum pairwise distance between objects in different clusters as the inter-cluster separation and the maximum diameter among all clusters as the intra-cluster compactness. The optimal cluster number isdetermined by maximizing the value of this index. To calculate the Dunn index, one has first to compute the distances between all the data points as follows: (13) where d(ci, cj) defines the inter-cluster distance between clusters Xi and Xj, d(Xk) is the intra-cluster distance of cluster (Xk), and c is the number of clusters [8486].

Visualisation using partial least squares regression.

Finally, we wish to analyse the relationship between automatically or manually determined trait categories and the geometric characteristics of the corresponding facial region. A common approach to establishing relationships in data is regression. The main difficulty in using dense data is the large number of correlated dependent variables in comparison to the number of observations, leading to model instability when using the linear least squares regression. We have addressed this problem by using the more advanced technique of partial least squares regression (PLSR) [41]

The aim of partial least squares regression is to establish a linear relationship between two sets of variables, X and Y, where X is the set of dependent variables and Y is the set of independent variables. In our case, Y is the set of dummy variables defining the manual or automatic categories, while X is the set of the x, y, and z coordinates of all vertices in a regularised facial mesh.

However, whilst categorical variables with two values may be directly entered as predictor or predicted variables in a multiple regression model, categorical variables with more than two values cannot be entered directly into a regression model. Therefore, we will use dummy variables [87]. A dummy variable is an artificial variable created to represent an attribute with two or more discrete values rather than continuous values as in standard regression. Therefore, dummy variables are created in such situations to force the regression algorithm to analyse variables correctly.

We converted the categorical variables (labels) into dummy variables [87]. For C categories, we need to C − 1 dummy variables before starting the regression process to determine their multiple and partial effects on the lip region.

The effects of trait categories on lip morphology can be illustrated using color maps. The regression coefficients define a set of weights at the vertices of the facial mesh. These weights define the magnitude and the direction of the vertex displacement per unit of the predictor (the predictors here are the label’s dummy variables). The values of interest represented in the heat maps are: (1) the ‘partial coefficients’ (magnitude); (2) the proportion of the variance at each vertex explained (partial R2) by the predictor; and (3) the degree of significance of the effect at each vertex [43, 44].

Statistical analysis.

In this research, analysis of variance (ANOVA) statistical method is used to test differences between two or more model or methods. ANOVA is a statistical analysis strategy for extraordinary tastefulness, utility and flexibility. It is the most effective technique accessible for breaking down the data from tests. ANOVA computer software is widely used for experimental tests, it is used to test general differences among model or methods [88]. Analysis of variance (ANOVA) is used to determine if the means of two or more groups are markedly different from each other. ANOVA checks the influence of one or more factors by comparing the means of different samples. The basic terminologies of ANOVA are:

  • Grand Mean: Mean is the average of a range of values. There are two types of means that are used in ANOVA test calculations, these are separate sample means and the grand mean. The grand mean is the mean of all observations combined.
  • Hypothesis: It is an educated guess about something in the world around us. It should be testable either by experiment or observation. ANOVA uses a Null hypothesis and an Alternate hypothesis. The Null hypothesis in ANOVA is correct if all the sample means are equal, or when they do not have significant difference. In contrast, the alternate hypothesis is correct if at least one of the sample means is different from the rest of the sample means.
  • Between Group Variability: It points to variations between the distributions of individual groups as the values within each group are different.
  • Within Group Variability: It points to variations caused by differences within individual groups. In other words, no interactions between group samples are considered [89, 90].

The Analysis of variance (ANOVA) test is simple applied test and it is used dramatically in many research for statistical analysis specialty in medical application and clinical diagnosis [89, 91, 92]. Therefore, in this research the ANOVA test is used to statistically analysis the effectiveness of geodesic curvature feature in lips morphology classification and categorisation.

Results

Six computational experiments were designed in order to assess the performance of proposed approaches for lip traits classification and categorisation. Experiments 1, 2, 3 and 4 were designed to investigate the best features to classify lip traits using the manual labels provided in [16]. In these experiments, the classification performance is measured using classification accuracy and AUC (Area Under ROC Curve) values. The ROC curve and AUC value are commonly used for evaluation local feature descriptors in both 2D and 3D images, for example in [93, 94]. That k-fold cross validation is a statistical method used to estimate the performance of machine learning algorithms.The choice of k is usually 5 or 10, but there is no formal rule. As k gets larger, the difference between the training set size and the testing subsets gets smaller [95].

In this work, The accuracy and AUC values are the average results for 5-folds (4745 ALSPAC face meshes were used in the classification task) cross validation runs. In Experiment 5, the proposed automatic approach for lip traits categorisation is assessed. Finally, in Experiment 6, the efficiency of the manual and automatic categorisation are compared using visualisation method. A detailed explanation of these experiments is provided in the following subsections.

Experiment 1: Classification based on 3D Euclidean distances

Euclidean distances are commonly used to study face morphology [31, 96, 97]. In this experiment, we used the inter-landmark Euclidean distances shown in Fig 8A to classify the lip traits. We combined several Euclidean distances defining a lip trait to form a classification descriptor; for example, five distances were combined together to classify the philtrum shape. Table 3 lists the accuracies as well as the AUC values for the classification performed using SVM and boosting classifiers for all lip traits except for the lower lip tone. Both the non-regularised and regularised meshes were considered. The accuracies suggest that Euclidean distance is a poor measure for the classification of lip traits. However, the results are approximately the same for both the non-regularised and regularised images, which indicates that Euclidean distance is not very sensitive to mesh resolution. S2 Table represents the Euclidean distance between the lips landmarks in Table 1.

thumbnail
Table 3. Classification results based on Euclidean distance.

https://doi.org/10.1371/journal.pone.0221197.t003

Experiment 2: Classification based on 3D geodesic distances

In the second experiment, we used geodesic distances to classify the lip traits. Many studies (e.g., see [98, 99]) suggest that geodesic distances could describe 3D models better than Euclidean distances. We calculated the geodesic distances using the fast marching and exact geodesic algorithms; these distances were between landmarks in the lip region, as shown in Fig 8A. Table 4 lists the classification results obtained with these features using, as above, two classification methods (SVM and boosting) for both the non-regularised and regularised meshes. The accuracies were slightly higher as compared to the results obtained using the respective Euclidean distances. The results also show little sensitivity to mesh resolution, since the exact method was utilised for much coarser, regularised meshes, thus providing higher performance than the fast marching method. S3 Table represents the geodesic distance between the lips landmarks in Table 1.

thumbnail
Table 4. Classification results based on geodesic distance.

https://doi.org/10.1371/journal.pone.0221197.t004

Experiment 3: Classification based on 3D geodesic path curvatures

The previous two experiments utilised the Euclidean and geodesic distances as morphological lip features, which are traditionally used for classification. In contrast, the third experiment is based on a geodesic path curvature descriptor. This descriptor combines the mean curvature, Gaussian curvature, shape index, and curvedness calculated for the points of the geodesic path between landmarks (Fig 8A). The curvatures of all geodesic paths defining a lip trait were combined together to form a feature descriptor of this trait for the classification purposes. S4, S5, S6, S7, S8, S9, S10 and S11 Tables record the mean curvature, Gaussian curvature, shape index, and curvedness for Cupid’s bow (cphL-ls and ls-cphR paths).

The lower lip tone was the most difficult morphological trait to classify, because its geometric features had to be determined in a narrow area very close to and below the lower lip contour. To this end, we used several geodesic paths approximately parallel to the lower lip contour to cover the desired area, as shown in Fig 8B. To find the optimal number of such paths and the distance between them, tuning tests were carried out followed by comparing the classification accuracies. The best accuracy was obtained for five paths separated in the vertical direction approximately by 2 mm. Table 5 lists the classification accuracies and AUC values for two classification methods (SVM and boosting) for the non-regularised and regularised meshes. The classification accuracies increased markedly, to 71–77% for the non-regularised and 67–72% for the regularised meshes. One can see that the classification accuracies depend on mesh resolution, because the curvature features are clearly sensitive to mesh resolution, which confirms the inference made in [100].

thumbnail
Table 5. Classification results based on geodesic path curvatures.

https://doi.org/10.1371/journal.pone.0221197.t005

Experiment 4: Classification based on a combination of features

In this experiment, the classification performance was explored by using a combination of the Euclidean distance, geodesic distances and geodesic curvature features as a classification descriptor. The Min-Max scaling approach was used to normalise the data to a fixed range of zero to one [101].

The best performance was achieved when the geodesic distance and curvature features were combined in the classification experiment. Table 6 displays the classification results based on different combinations of features for the non-regularised and regularised faces. Using a descriptor combining the geodesic curvature features and geodesic distances increased the classification accuracy to 72–79% for the non-regularised meshes and 70–74% for the regularised meshes.

thumbnail
Table 6. Classification results based on different combinations of features.

https://doi.org/10.1371/journal.pone.0221197.t006

Experiment 5: Lip traits automatic categorisation

In this experiment, we used the combination of the above geodesic curvatures and the geodesic distances to categorise (cluster) the lip morphological traits without relying on manual labeling. An unsupervised clustering scheme was utilised to partition the geodesic curvature features into multiple clusters, each defined by a centroid, and the Kmeans++ algorithm was employed to perform the clustering.

Kmeans++ starts with allocation one cluster center randomly and then searches for other centers given the first one. So this algorithm uses random initialization as a starting point; hence, it can give different results on different runs. Therefore, the Kmeans++ clustering procedure was run 100 times to reduce the effect of randomisation. Out of 100 results, the clustering results that produced the minimal sum of squared distance scores was chosen. An analysis for optimum numbers of clusters was carried out using internal cluster validation techniques. We computed the validity indexes CH, DI, and SI for all traits with the number of clusters K, ranging from 2 to 9, with the tests repeated 50 times to find a stable number of clusters and labels. For example, Fig 9 displays the results for the philtrum shape; the optimum number of clusters is 5 (based on DI and SI) or 7 (based on CH). For the other traits, the optimum number of clusters was obtained in the same manner: 3 for Cupid’s bow, 3 or 4 for the upper lip contour, 3 or 4 for the lower lip contour, 5 for the lower lip-chin shape, and 5 for the lower lip tone shape. These results are almost the same as those produced by a medical expert in [16]. Although there are slight differences: for example, the philtrum shape was categorised subjectively into 7 clusters, while automatic categorisation produced five clusters on some runs and seven clusters on other runs. To compare with the manual results in the next experiment, the philtrum shape with 7 automatic clusters was used. Table 7 shows the percentage of the number of times the validation methods chose a certain number of clusters. While Table 8 records the number of samples in each sub class of lip traits which automatically categorised.

thumbnail
Fig 9. Validity indices for the number of clusters for the philtrum shape.

The figure shows the CH, DI, and SI results.

https://doi.org/10.1371/journal.pone.0221197.g009

thumbnail
Table 7. The percentage of the number of times the validation methods chose a certain number of clusters.

https://doi.org/10.1371/journal.pone.0221197.t007

The best classification accuracy for the manual labels was in the range of 72–79% (Table 6). The same classification procedure was repeated now with the automatic labels. Table 9 shows the classification accuracies for a descriptor combining the geodesic curvature features and geodesic distances, obtained using both manual and automatic lip morphological traits labels. In addition, the lower lip tone classification accuracy was as high as 92.7% with the automatic categorisation labels and geodesic curvatures features. However, for the manual categorisation labels and the same features, the classification accuracy for this trait was only 76.8%. It is clear that the classification based on the automatic labels outperforms that based on the manual labels.

thumbnail
Table 9. Classification accuracies for the manual and automatic lip area trait labels.

https://doi.org/10.1371/journal.pone.0221197.t009

Experiment 6: Visualisation of the effect of trait categories on the lip region

PLS regression was used to characterise the effects of the trait categories on the regularised 3D faces. All statistical significance tests were based on 1000 permutations [102]. The partial effects (one variable is independent of the others) in the multivariate regression were coded by the partial regression coefficients. These coefficients define label weights at the mesh vertices, which were visualised as a heat map; cooler colours correspond to weaker effects on vertices, while warmer colours correspond to stronger effects.

As an example, Fig 10 illustrates the regression results for the manual philtrum shape labels, while Fig 11 shows those for the automatic philtrum shape labels. In these figures, the “partial coefficients” correspond to the magnitude of the vertex displacement in three dimensions; the proportion of variance that the predictor variable predicts at each individual vertex is defined by partial R2. The effect of the labels was displayed as colour maps of statistical significance using two colours, with yellow indicating highly significant results (p-value < 0.001) and green showing moderately significant or insignificant results (p-value ≥ 0.001).

thumbnail
Fig 10. Regression results.

Effect of a manual label (philtrum shape) on the lower face. (A) Partial coefficients. (B) R2. (C) p < 0.001 (yellow), p ≥ 0.001 (green). Warmer colours correspond to stronger effects.

https://doi.org/10.1371/journal.pone.0221197.g010

thumbnail
Fig 11. Regression results.

Effect of an automatic label (philtrum shape) on the lower face. (A) Partial coefficients. (B) R2. (C) p < 0.001 (yellow), p ≥ 0.001 (green). Warmer colours correspond to stronger effects.

https://doi.org/10.1371/journal.pone.0221197.g011

Figs 12 and 13 visualise the effects for all trait’s dummy variables as individual and multiple (the combination of all trait’s dummy variables). As can be seen from the visualisation, the multiple effect of all trait categories is concentrated and significant in the lip region.

thumbnail
Fig 12. Visualisation of the effect of manual categories on the lower face based on the regression results for dummy variables.

The ‘Partial coeffs’ columns display heat maps of the partial regression coefficients associated with mesh vertices (warmer colours correspond to stronger effects). The ‘R2’ columns display heat maps of proportion of the variance. The ‘p < 0.001’ columns show two-colour maps of the statistical significance of the effect: yellow for p-value < 0.001 and green for p-value ≥ 0.001.

https://doi.org/10.1371/journal.pone.0221197.g012

thumbnail
Fig 13. Visualisation of the effect of automatic categories on the lower face based on the regression results for dummy variables.

The ‘Partial coeffs’ columns display heat maps of the partial regression coefficients associated with mesh vertices (warmer colours correspond to stronger effects). The ‘R2’ columns display heat maps of proportion of the variance. The ‘p < 0.001’ columns show two-colour maps of the statistical significance of the effect: yellow for p-value < 0.001 and green for p-value ≥ 0.001.

https://doi.org/10.1371/journal.pone.0221197.g013

ANOVA test results

To determine the significance of classification improvement using the geodesic curvature features over the geodesic distance and euclidean distance features, the results of ten runs of the Boosting classifier using these three types of geometric features (Euclidean distance, geodesic distance and geodesic curvature features) are submitted to Analysis of Variance (ANOVA) statistical testing. Table 10 shows the ANOVA results, where P-value is the probability of the improvement to occur by chance, and MS is the mean square error.The F-value is a ratio Between group variability to Within group variability. The improvement is significant if the P-value is less than 0.05, which means that the improvement is unlikely to happen by chance. As the P-value is almost zero in Geodesic curvature features cases, the improvement of the proposed features over the geodesic distance and Euclidean distance features is significant.

Discussion

The first three experiments in the present study aimed to determine which facial features were the most effective in automatic classification of the lip traits using non-regularised and regularised 3D meshes. Although fairly effective for both high and low resolution data, the Euclidean distances alone did not produce good enough classification accuracies. Experiment 2 showed that the geodesic distances between landmarks provided higher classification accuracies. We attribute this to geodesic distances being more informative than Euclidean distances in describing facial surfaces; this compatible with many state of the art research such as [103, 104]. The geodesic distance features classification accuracies are higher than Euclideann distance features classification accuracies and with only small differences between the non-regularised and regularised data. Experiment 3 used 3D geometric curvatures of the shortest geodesic path between two anthropometric landmarks as features in the classification experiments. The accuracies were found to improve markedly, which is likely due to the curvature features taking into account local facial geometry and, hence, characterising the lip shapes much better than the Euclidean and geodesic distances.

Most of the previous studies [33, 98, 105] have focused on using combinations of facial features to achieve higher classification accuracies. In the present study, we also adopted this approach and used a combination of Euclidean and geodesic distances with geodesic curvature features (see Table 6) to produce feature descriptors for the lip morphology classification. The highest accuracy was obtained when geodesic distances were combined with geodesic curvatures. The SVM (Support Vector Machine) method failed to classify the lip traits efficiently, in contrast to the boosting method, because the data were highly imbalanced in more than one class; this finding seems to corroborate the inference made in [106]. In spite of using different types and combination of features to classify the lip traits, the classification accuracies were not very high for the manual labels provided in [16]. This encouraged us to categorise the lip traits automatically. The Kmeans++ algorithm has shown good categorisation results in our experiments, which, however, does not prevent us from trying alternative clustering techniques, such as spectral clustering [107], or using distance measures to cluster the elements of classes [108] in the future. Three internal validation techniques were used to select an optimum number of clusters for each lip trait. The selection process was repeated 50 times to find stable numbers of clusters. For the lower and upper lip contours, these techniques were found to show quite contradicting results for three and four clusters. For this reason, we used the labeling results for both alternatives in Experiment 6.

Table 10 records the Analysis of Variance (ANOVA) test results. As the P-value is almost zero in Geodesic curvature features cases, the improvement of the proposed features over the geodesic distance and Euclidean distance features is significant. Using the combination features (geodesic distance and geodesic curvature features) provide smallest p-value but this value is not far from the p-value in geodesic curvature features case. From that, the ability of geodesic curvature feature in lips morpholoy is proved.

In Experiment 6, a new method was used to visualise the automatic and subjective (manual) categorisation (labeling) regression results for the lip area. All previous studies [42, 44], and [43] dealt with continuous variables. By contrast, the present study seems to be the first to deal with discrete variables (automatic or subjective (manual) labels) for face morphology classification. Using dummy variables is a way to deal with the discrete variable problem; for example, six dummy variables were used to represent the philtrum shape labels which fall into seven classes. Figs 10 and 11 show the regression results for the philtrum shape trait with one dummy variable.

Fig 12 details the regression results for the manual labels showing the multivariate effect and contribution of each individual dummy variable. Fig 13 details the regression results for the automatic labels.

On careful examination of the results, one can see that our automatic labels (categories) are fairly similar to the subjective labels (categories) [16]. However, for three out of seven traits (the lower lip-chin, the lower lip contour and lower lip tone shape), our automatic labels provide much better categorisation results, as the label effect appears at the right areas (compare Figs 12 and 13). In particular, for the lower lip tone, the subjective labels show a strong effect (red colours) near the oral commissures rather than in the mentolabial sulcus area, which is under the lower lip contour. The analysis does not give us a clear indication of the optimum number of clusters for the classification of the upper and lower lip contours. However, it is apparent from Fig 13 that using four clusters would be preferable in both cases as the colour maps for respective dummy variables highlight the correct areas better. The classification accuracy using the automatic categories has outperformed the manual categories classification accuracy by at least 8%. All this testifies that the approach for automatic categorisation of 3D facial morphology proposed in this study has a considerable potential,this categorisation technique and geodesic curvature features can be used for different facial trait to save clinicians some manual labor and produce accurate facial morphological traits categorisation results in a short time comparing to manual work.

Conclusion

We proposed a new automatic approach to classify and categorise various facial morphological traits, with a specific application to morphological traits in the lip region, using 3D geometric features based on geodesic path curvatures. We evaluated our approach on the large ALSPAC dataset consisting of 4747 3D scans of a 15-year-old population. Six experiments were conducted to evaluate our classification and categorisation approach and to compare manual and automatic trait categories on the lip region models using the PLSR method.

In general, the geodesic curvature features provided higher classification accuracies as compared to the Euclidean and geodesic distances. The classification accuracies increased when both geodesic distance and geodesic curvature were used. The same features were used to categorise lip traits automatically, with categorisation approach based on the Kmeans++ and internal cluster validation algorithms. The results of Experiment 6 illustrate that the automatic categories are more accurate in defining lip traits as compared to the manual categories.

In future, we are planning to extend our approach to other morphological facial features such as the nose. The proposed approach may be have potential for gaining knowledge about genotype and facial traits associations, which will also be considered in future work.

Acknowledgments

We are extremely grateful to all of the families who took part in this study, the midwives for their help in recruiting them, and the whole ALSPAC team, which includes interviewers, computer and laboratory technicians, clerical workers, research scientists, volunteers, managers, receptionists, and nurses. This publication is the work of Hawraa Abbas and the other co-authors who will serve as guarantors of the contents of this paper.

References

  1. 1. Solow B, Tallgren A. Natural head position in standing subjects. Acta Odontologica Scandinavica. 1971;29(5):591–607. pmid:5290983
  2. 2. Sforza C, Grandi G, Binelli M, Dolci C, De Menezes M, Ferrario VF. Age- and sex-related changes in three-dimensional lip morphology. Forensic science international. 2010;200(1):182–e1. pmid:20570070
  3. 3. De Menezes M, Rosati R, Baga I, Mapelli A, Sforza C. Three-dimensional analysis of labial morphology: effect of sex and age. International journal of oral and maxillofacial surgery. 2011;40(8):856–861. pmid:21477995
  4. 4. Shelke RG, Annadate S. Face Recognition and Gender Classification Using Feature of Lips. International Journal of Innovation and Scientific Research (IJISR), Innovative Space of Scientific Research Journals (ISSR). 2014;10(2).
  5. 5. Stewart D, Pass A, Zhang J. Gender classification via lips: static and dynamic features. IET biometrics. 2013;2(1):28–34.
  6. 6. Koller O, Ney H, Bowden R. Deep Learning of Mouth Shapes for Sign Language. In: Proceedings of the IEEE International Conference on Computer Vision Workshops; 2015. p. 85–91.
  7. 7. Saitoh T, Hisagi M, Konishi R. Japanese Phone Recognition Using Lip Image Information. In: MVA. Citeseer; 2007. p. 134–137.
  8. 8. Holdaway RA. A soft-tissue cephalometric analysis and its use in orthodontic treatment planning. Part I. American journal of orthodontics. 1983;84(1):1–28. pmid:6575614
  9. 9. Merrifield LL. The profile line as an aid in critically evaluating facial esthetics. American journal of orthodontics. 1966;52(11):804–822. pmid:5223046
  10. 10. Ricketts RM. Esthetics, environment, and the law of lip relation. American journal of orthodontics. 1968;54(4):272–289. pmid:5238879
  11. 11. Burstone CJ. Lip posture and its significance in treatment planning. American journal of orthodontics. 1967;53(4):262–284. pmid:5227460
  12. 12. Wang KH, Heike CL, Clarkson MD, Mejino JL, Brinkley JF, Tse RW, et al. Evaluation and integration of disparate classification systems for clefts of the lip. Frontiers in physiology. 2014;5:163. pmid:24860508
  13. 13. Beaty T, Taub M, Scott A, Murray J, Marazita M, Schwender H, et al. Confirming genes influencing risk to cleft lip with/without cleft palate in a case–parent trio study. Human genetics. 2013;132(7):771–781.
  14. 14. Suttie M, Foroud T, Wetherill L, Jacobson JL, Molteno CD, Meintjes EM, et al. Facial dysmorphism across the fetal alcohol spectrum. Pediatrics. 2013;131(3):e779–e788. pmid:23439907
  15. 15. Abbas H, Hicks Y, Marshall D, Zhurov A, Richmond S. A 3D morphometric perspective for facial gender analysis and classification using geodesic path curvature features. Computational Visual Media. 2017.
  16. 16. Wilson C, Playle R, Toma A, Zhurov A, Ness A, Richmond S. The prevalence of lip vermilion morphological traits in a 15-year-old population. American Journal of Medical Genetics Part A. 2013;161(1):4–12.
  17. 17. Hwang K, Kim DJ, Hwang SH. Musculature of the pars marginalis of the upper orbicularis oris muscle. Journal of Craniofacial Surgery. 2007;18(1):151–154. pmid:17251855
  18. 18. Heidari Z, Mahmoudzadeh-Sagheb H, Rad AA, Dahmardeh N. Anthropometric measurements of the lips in 18–25-year-old men of Sistani and Baluch descent. Bull Env Pharmacol Life Sci. 2014;3:139–142.
  19. 19. Carey JC, Cohen MM, Curry CJ, Devriendt K, Holmes LB, Verloes A. Elements of morphology: standard terminology for the lips, mouth, and oral region. American Journal of Medical Genetics Part A. 2009;149(1):77–92.
  20. 20. Shi JY, Zhou H, Mao RY, Chen Y, Li JT, Huo HY. A preliminary study on the key factors contributing to the attractive lips of Chinese children. Asian Pacific journal of tropical medicine. 2012;5(4):318–322. pmid:22449526
  21. 21. O’Toole AJ, Vetter T, Troje NF, Bülthoff HH. Sex classification is better with three-dimensional head structure than with image intensity information. Perception. 1997;26(1):75–84. pmid:9196691
  22. 22. Bruce V, Young A. Understanding face recognition. British journal of psychology. 1986;77(3):305–327. pmid:3756376
  23. 23. Mori A, Nakajima T, Kaneko T, Sakuma H, Aoki Y. Analysis of 109 Japanese children’s lip and nose shapes using 3-dimensional digitizer. British journal of plastic surgery. 2005;58(3):318–329.
  24. 24. Lee J, Ku B, Combs PD, Da Silveira AC, Markey MK. Quantitative Anthropometric Measures of Facial Appearance of Healthy Hispanic/Latino White Children: Establishing Reference Data for Care of Cleft Lip With or Without Cleft Palate. 3D Research. 2017;8(2):19.
  25. 25. Alpaydin E. Introduction to machine learning. MIT press; 2014.
  26. 26. Caruana R, Niculescu-Mizil A. An empirical comparison of supervised learning algorithms. In: Proceedings of the 23rd international conference on Machine learning. ACM; 2006. p. 161–168.
  27. 27. Enver K, Erdo P, Polat K. Brain MRI segmentation based on different clustering algorithms. International Journal of Computer Applications. 2016;155(3):37–40.
  28. 28. Anthony G, Gregg H, Tshilidzi M. Image classification using SVMs: one-against-one vs one-against-all. arXiv preprint arXiv:07112914. 2007.
  29. 29. Yuan X, Abouelenien M. A multi-class boosting method for learning from imbalanced data. International Journal of Granular Computing, Rough Sets and Intelligent Systems. 2015;4(1):13–29.
  30. 30. Arthur D, Vassilvitskii S. k-means++: The advantages of careful seeding. In: Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms. Society for Industrial and Applied Mathematics; 2007. p. 1027–1035.
  31. 31. Toma AM. Characterization of normal facial features and their association with genes [PhD Thesis]. Cardiff University; 2014.
  32. 32. Wilamowska K, Wu J, Heike C, Shapiro L. Shape-based classification of 3D facial data to support 22q11.2DS craniofacial research. Journal of digital imaging. 2012;25(3):400–408. pmid:22086243
  33. 33. Abbas H, Hicks Y, Marshall D. Automatic classification of facial morphology for medical applications. Procedia Computer Science. 2015;60:1649–1658.
  34. 34. Perakis P, Theoharis T, Passalis G, Kakadiaris IA. Automatic 3D Facial Region Retrieval from Multi-pose Facial Datasets. In: 3DOR; 2009. p. 37–44.
  35. 35. Lu X, Jain AK. Automatic feature extraction for multiview 3D face recognition. In: Automatic Face and Gesture Recognition, 2006. FGR 2006. 7th International Conference on. IEEE; 2006. p. 585–590.
  36. 36. Lu X, Jain AK, Colbry D. Matching 2.5D face scans to 3D models. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2006;28(1):31–43.
  37. 37. Zhao J, Liu C, Wu Z, Duan F, Zhang M, Wang K, et al. 3D facial similarity measure based on geodesic network and curvatures. Mathematical Problems in Engineering. 2014;2014.
  38. 38. Ballihi L, Amor BB, Daoudi M, Srivastava A, Aboutajdine D. Boosting 3-D-geometric features for efficient face recognition and gender classification. IEEE Transactions on Information Forensics and Security. 2012;7(6):1766–1779.
  39. 39. Jahanbin S, Choi H, Liu Y, Bovik AC. Three dimensional face recognition using iso-geodesic and iso-depth curves. In: Biometrics: Theory, Applications and Systems, 2008. BTAS 2008. 2nd IEEE International Conference on. IEEE; 2008. p. 1–6.
  40. 40. Ahdid R, Barrah EM, Safi S, Manaut B. Facial Surface Analysis using Iso-Geodesic Curves in Three Dimensional Face Recognition System. arXiv preprint arXiv:160808878. 2016.
  41. 41. com S. Partial Least Squares (PLS);. Available from: http://www.statsoft.com/Textbook/Partial-Least-Squares.
  42. 42. Shrimpton S, Daniels K, De Greef S, Tilotta F, Willems G, Vandermeulen D, et al. A spatially-dense regression study of facial form and tissue depth: towards an interactive tool for craniofacial reconstruction. Forensic science international. 2014;234:103–110. pmid:24378309
  43. 43. Claes P, Daniels K, Vandermeulen D, Suetens P, Shriver MD. A PLS regression framework for spatially-dense geometric morphometrics to analyze effects on shape and shape characteristics: applied to the study of genomic ancestry and sex on facial morphology. In: BIOLOGICAL SHAPE ANALYSIS: Proceedings of the 3rd International Symposium; 2015. p. 205–215.
  44. 44. Matthews H, Penington T, Saey I, Halliday J, Muggli E, Claes P. Spatially dense morphometrics of craniofacial sexual dimorphism in 1-year-olds. Journal of Anatomy. 2016;229(4):549–559. pmid:27338586
  45. 45. Boyd A, Golding J, Macleod J, Lawlor DA, Fraser A, Henderson J, et al. Cohort profile: the ‘children of the 90s’—the index offspring of the Avon Longitudinal Study of Parents and Children. International journal of epidemiology. 2013;42(1):111–127.
  46. 46. Kau CH, Richmond S. Three-dimensional analysis of facial morphology surface changes in untreated children from 12 to 14 years of age. American Journal of Orthodontics and Dentofacial Orthopedics. 2008;134(6):751–760. pmid:19061801
  47. 47. Toma AM, Zhurov A, Playle R, Ong E, Richmond S. Reproducibility of facial soft tissue landmarks on 3D laser-scanned facial images. Orthodontics & craniofacial research. 2009;12(1):33–42.
  48. 48. Ruiz M, Illingworth J. Automatic landmarking of faces in 3D-ALF3D. In: Visual Information Engineering, 2008. VIE 2008. 5th International Conference on. IET; 2008. p. 41–46.
  49. 49. Koppen WP. Learning 3D face shape features from local coherence [PhD Thesis]. Centre for Vision, Speech and Signal Processing Faculty of Electronic Engineering University of Surrey; 2014.
  50. 50. de Jong MA, Wollstein A, Ruff C, Dunaway D, Hysi P, Spector T, et al. An automatic 3D facial landmarking algorithm using 2D Gabor wavelets. IEEE Transactions on Image Processing. 2016;25(2):580–588. pmid:26540684
  51. 51. Zhurov A, Richmond S, Kau CH, Toma A. Averaging facial images. In: Kau CH, Richmond S, editors. Three-dimensional imaging for orthodontics and maxillofacial surgery. Wiley-Blackwell; 2010. p. 126–144.
  52. 52. Hammond P, Suttie M, Hennekam RC, Allanson J, Shore EM, Kaplan FS. The face signature of fibrodysplasia ossificans progressiva. American Journal of Medical Genetics Part A. 2012;158(6):1368–1380.
  53. 53. Claes P, Walters M, Clement J. Improved facial outcome assessment using a 3D anthropometric mask. International journal of oral and maxillofacial surgery. 2012;41(3):324–330. pmid:22103995
  54. 54. Zhurov AI, Kau CH, Richmond S. Computer methods for measuring 3D facial morphology. In: Middleton J, Shrive N, Jones M, editors. CMBBE2004 Proceedings. FIRST Numerics Ltd.; 2005.
  55. 55. Sorkine O, Cohen-Or D, Lipman Y, Alexa M, Rössl C, Seidel HP. Laplacian surface editing. In: Proceedings of the 2004 Eurographics/ACM SIGGRAPH symposium on Geometry processing. ACM; 2004. p. 175–184.
  56. 56. Drira H, Amor BB, Daoudi M, Srivastava A. Pose and expression-invariant 3d face recognition using elastic radial curves. In: British machine vision conference; 2010. p. 1–11.
  57. 57. Peyre G. Toolbox graph—A toolbox to process graph and triangulated meshes; Updated 27 Jun 2009. Available from: https://uk.mathworks.com/matlabcentral/fileexchange/5355-toolbox-graph/content/toolbox_graph/html/content.html.
  58. 58. Szeptycki P, Ardabilian M, Chen L. A coarse-to-fine curvature analysis-based rotation invariant 3D face landmarking. In: 2009 IEEE 3rd International Conference on Biometrics: Theory, Applications, and Systems. IEEE; 2009. p. 1–6.
  59. 59. Toma AM, Zhurov AI, Playle R, Marshall D, Rosin PR, Richmond S. The assessment of facial variation in 4747 British school children. European Journal of Orthodontics. 2012;34:665–664.
  60. 60. Besl PJ, McKay ND. Method for registration of 3-D shapes. In: Robotics-DL tentative. International Society for Optics and Photonics; 1992. p. 586–606.
  61. 61. Dudani SA. The distance-weighted k-nearest-neighbor rule. IEEE Transactions on Systems, Man, and Cybernetics. 1976;SMC-6(4):325–327.
  62. 62. Djordjevic J, Zhurov AI, Richmond S, Consortium V, et al. Genetic and environmental contributions to facial morphological variation: a 3D population-based twin study. PloS one. 2016;11(9):e0162250. pmid:27584156
  63. 63. Gilani SZ, Mian A. Perceptual differences between men and women: A 3D facial morphometric perspective. In: Pattern Recognition (ICPR), 2014 22nd International Conference on. IEEE; 2014. p. 2413–2418.
  64. 64. Mpiperis I, Malassiotis S, Strintzis MG. 3-D Face Recognition With the Geodesic Polar Representation. IEEE Transactions on Information Forensics and Security. 2007;2(3-2):537–547.
  65. 65. Sethian JA. Level set methods and fast marching methods: evolving interfaces in computational geometry, fluid mechanics, computer vision, and materials science. vol. 3. Cambridge university press; 1999.
  66. 66. Surazhsky V, Surazhsky T, Kirsanov D, Gortler SJ, Hoppe H; ACM. Fast exact and approximate geodesics on meshes. ACM transactions on graphics (TOG). 2005;24(3):553–560.
  67. 67. Xin SQ, Wang GJ. Improving Chen and Han’s algorithm on the discrete geodesic problem. ACM Transactions on Graphics (TOG). 2009;28(4):104.
  68. 68. Weber O, Devir YS, Bronstein AM, Bronstein MM, Kimmel R. Parallel algorithms for approximation of distance maps on parametric surfaces. ACM Transactions on Graphics (TOG). 2008;27(4):104.
  69. 69. Peyre G. Fast marching MATLAB toolbox; Updated 19 Jul 2009. Available from: https://uk.mathworks.com/matlabcentral/fileexchange/6110-toolbox-fast-marching.
  70. 70. Kirsanov D. Exact geodesic for triangular meshes; Updated 03 Mar 2008. Available from: https://uk.mathworks.com/matlabcentral/fileexchange/18168-exact-geodesic-for-triangular-meshes.
  71. 71. Bommes D, Kobbelt L. Accurate Computation of Geodesic Distance Fields for Polygonal Curves on Triangle Meshes. In: VMV. vol. 7; 2007. p. 151–160.
  72. 72. Meyer M, Desbrun M, Schröder P, Barr AH, et al. Discrete differential-geometry operators for triangulated 2-manifolds. Visualization and mathematics. 2002;3(2):52–58.
  73. 73. Goldfeather J, Interrante V. A novel cubic-order algorithm for approximating principal direction vectors. ACM Transactions on Graphics (TOG). 2004;23(1):45–63.
  74. 74. Fu JH, et al. Convergence of curvatures in secant approximations. Journal of Differential Geometry. 1993;37(1):177–190.
  75. 75. Alliez P, Cohen-Steiner D, Devillers O, Lévy B, Desbrun M; ACM. Anisotropic polygonal remeshing. ACM Transactions on Graphics (TOG). 2003;22(3):485–493.
  76. 76. Abbas H, Hicks Y, Marshall D, Zhurov AI, Richmond S. A 3D morphometric perspective for facial gender analysis and classification using geodesic path curvature features. Computational Visual Media. 2018; p. 1–16.
  77. 77. Howitt D, Cramer D. Introduction to statistics in psychology. Pearson education; 2007.
  78. 78. Law MT, Thome N, Cord M. Bag-of-words image representation: Key ideas and further insight. In: Fusion in Computer Vision. Springer; 2014. p. 29–52.
  79. 79. Ramyachitra D, Manikandan P. Imbalanced Dataset Classification and Solutions: A Review. International Journal of Computing and Business Research. 2014;5(4).
  80. 80. Burns N, Bi Y, Wang H, Anderson T. Sentiment analysis of customer reviews: Balanced versus unbalanced datasets. Knowledge-Based and Intelligent Information and Engineering Systems. 2011; p. 161–170.
  81. 81. Satyasree K, Murthy J. An exhaustive literature review on class imbalance problem. Int J Emerg Trends Technol Comput Sci. 2013;2:109–118.
  82. 82. Paris S. Multiclass GentleAdaboosting; Updated 24 Nov 2011. Available from: https://uk.mathworks.com/matlabcentral/fileexchange/22997-multiclass-gentleadaboosting.
  83. 83. Khan F. An initial seed selection algorithm for k-means clustering of georeferenced data to improve replicability of cluster assignments for mapping application. Applied Soft Computing. 2012;12(11):3698–3700.
  84. 84. Rendón E, Abundez I, Arizmendi A, Quiroz E. Internal versus external cluster validation indexes. International Journal of computers and communications. 2011;5(1):27–34.
  85. 85. Rendón E, Abundez IM, Gutierrez C, Zagal SD, Arizmendi A, Quiroz EM, et al. A comparison of internal and external cluster validation indexes. In: Proceedings of the 2011 American conference on applied mathematics and the 5th WSEAS international conference on computer engineering and applications, San Francisco, CA, USA. vol. 29; 2011. p. 158–163.
  86. 86. Liu Y, Li Z, Xiong H, Gao X, Wu J. Understanding of internal clustering validation measures. In: 2010 IEEE International Conference on Data Mining. IEEE; 2010. p. 911–916.
  87. 87. Stockburger DW. Multiple Regression with Categorical Variables;. Available from: http://www.psychstat.missouristate.edu/multibook/mlt08m.html.
  88. 88. Algina J, Olejnik S. Conducting power analyses for ANOVA and ANCOVA in between-subjects designs. Evaluation & the Health Professions. 2003;26(3):288–314.
  89. 89. Armstrong RA, Slade S, Eperjesi F. An introduction to analysis of variance (ANOVA) with special reference to data from clinical experiments in optometry. Ophthalmic and Physiological Optics. 2000;20(3):235–241.pmid:10897345
  90. 90. Chatfield C. Introduction to multivariate analysis. Routledge; 2018.
  91. 91. Serrano C, García-Fernández L, Fernández-Blázquez JP, Barbeck M, Ghanaati S, Unger R, et al. Nanostructured medical sutures with antibacterial properties. Biomaterials. 2015;52:291–300. pmid:25818435
  92. 92. Shrestha BK, Mousa HM, Tiwari AP, Ko SW, Park CH, Kim CS. Development of polyamide-6, 6/chitosan electrospun hybrid nanofibrous scaffolds for tissue engineering application. Carbohydrate polymers. 2016;148:107–114. pmid:27185121
  93. 93. Tombari F, Salti S, Di Stefano L. Unique signatures of histograms for local surface description. In: European conference on computer vision. Springer; 2010. p. 356–369.
  94. 94. Guo Y, Bennamoun M, Sohel F, Lu M, Wan J, Kwok NM. A comprehensive performance evaluation of 3D local feature descriptors. International Journal of Computer Vision. 2016;116(1):66–89.
  95. 95. Kuhn M, Johnson K. Applied predictive modeling. vol. 26. Springer; 2013.
  96. 96. Burton AM, Bruce V, Dench N. What’s the difference between men and women? Evidence from facial measurement. Perception. 1993;22(2):153–176.
  97. 97. Seo H, Song Y, Kim C, Kim H. Characteristics of Korean Children’s Facial Anthropometry Evaluated by Three-dimensional Imaging. Journal of the International Society for Respiratory Protection Vol. 2016;33(1).
  98. 98. Gupta S, Markey MK, Bovik AC. Anthropometric 3D face recognition. International journal of computer vision. 2010;90(3):331–349.
  99. 99. Hamza AB, Krim H. Geodesic matching of triangulated surfaces. Image Processing, IEEE Transactions on. 2006;15(8):2249–2258.
  100. 100. Gatzke T, Grimm C. Improved curvature estimation on triangular meshes. Computer Science and Engineering, Washington University in St. Louis; 2004. WUCSE-2004–9. Available from: https://openscholarship.wustl.edu/cse_research/1056/.
  101. 101. Raschka S. About Feature Scaling and Normalization; Jul 11, 2014. Available from: https://sebastianraschka.com/Articles/2014_about_feature_scaling.html#about-standardization.
  102. 102. Anderson MJ, Legendre P. An empirical comparison of permutation methods for tests of partial regression coefficients in a linear model. Journal of statistical computation and simulation. 1999;62(3):271–303.
  103. 103. Bronstein AM, Bronstein MM, Kimmel R. Efficient computation of isometry-invariant distances between surfaces. SIAM Journal on Scientific Computing. 2006;28(5):1812–1836.
  104. 104. Vezzetti E, Marcolin F. 3D human face description: landmarks measures and geometrical features. Image and Vision Computing. 2012;30(10):698–712.
  105. 105. Gilani SZ, Rooney K, Shafait F, Walters M, Mian A. Geometric facial gender scoring: objectivity of perception. PloS one. 2014;9(6):e99483. pmid:24923319
  106. 106. Brown I, Mues C. An experimental comparison of classification algorithms for imbalanced credit scoring data sets. Expert Systems with Applications. 2012;39(3):3446–3453.
  107. 107. Ng AY, Jordan MI, Weiss Y, et al. On spectral clustering: Analysis and an algorithm. Advances in neural information processing systems. 2002;2:849–856.
  108. 108. Galluccio L, Michel O, Comon P, Kliger M, Hero AO. Clustering with a new distance measure based on a dual-rooted tree. Information Sciences. 2013;251:96–113.