<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">IASC</journal-id>
<journal-id journal-id-type="nlm-ta">IASC</journal-id>
<journal-id journal-id-type="publisher-id">IASC</journal-id>
<journal-title-group>
<journal-title>Intelligent Automation &#x0026; Soft Computing</journal-title>
</journal-title-group>
<issn pub-type="epub">2326-005X</issn><issn pub-type="ppub">1079-8587</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">19533</article-id>
<article-id pub-id-type="doi">10.32604/iasc.2022.019533</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Semantic Human Face Analysis for Multi-level Age Estimation</article-title><alt-title alt-title-type="left-running-head">Semantic Human Face Analysis for Multi-level Age Estimation</alt-title><alt-title alt-title-type="right-running-head">Semantic Human Face Analysis for Multi-level Age Estimation</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author" corresp="yes">
<name name-style="western">
<surname>Howyan</surname>
<given-names>Rawan Sulaiman</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
<xref ref-type="aff" rid="aff-2">2</xref>
<email>rhowian@stu.kau.edu.sa</email>
</contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western">
<surname>Jaha</surname>
<given-names>Emad Sami</given-names>
</name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<aff id="aff-1">
<label>1</label><institution>Department of Computer Science, Faculty of Computing and Information Technology, King Abdulaziz University</institution>, <addr-line>Jeddah</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-2">
<label>2</label><institution>Department of Computer Science, Faculty of Computers and Information Technology, University of Tabuk</institution>, <addr-line>Tabuk</addr-line>, <country>Saudi Arabia</country></aff>
</contrib-group><author-notes><corresp id="cor1">&#x002A;Corresponding Author: Rawan Sulaiman Howyan. Email: <email>rhowian@stu.kau.edu.sa</email> or <email>rhowian@ut.edu.sa</email></corresp></author-notes>
<pub-date pub-type="epub" date-type="pub" iso-8601-date="2021-08-21">
<day>21</day>
<month>8</month>
<year>2021</year>
</pub-date>
<volume>31</volume>
<issue>1</issue>
<fpage>555</fpage>
<lpage>580</lpage>
<history>
<date date-type="received">
<day>16</day>
<month>4</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>05</day>
<month>6</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2021 Howyan and Jaha</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Howyan and Jaha</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_IASC_19533.pdf"></self-uri>
<abstract>
<p>Human face is one of the most widely used biometrics based on computer-vision to derive various useful information such as gender, ethnicity, age, and even identity. Facial age estimation has received great attention during the last decades because of its influence in many applications, like face recognition and verification, which may be affected by aging changes and signs which appear on human face along with age progression. Thus, it becomes a prominent challenge for many researchers. One of the most influential factors on age estimation is the type of features used in the model training process. Computer-vision is characterized by its superior ability to extract traditional facial features such as shape, size, texture, and deep features. However, it is still difficult for computers to extract and deal with semantic features inferred by human-vision. Therefore, we need somehow to bridge the semantic gap between machines and humans to enable utilization of the human brain capabilities of perceiving and processing visual information in semantic space. Our research aims to exploit human-vision in semantic facial feature extraction and fusion with traditional computer-vision features to obtain integrated and more informative features as an initial study paving the way to further augment the outperforming state-of-the-art age estimation models. A hierarchical automatic age estimation is achieved upon two consecutive stages: classification to predict (high-level) age group, followed by regression to estimate (low-level) exact age. The results showed noticeable performance improvements, when fusing semantic-based features with traditional vision-based features, surpassing the performance of traditional features alone.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Semantic features</kwd>
<kwd>semantic face analysis</kwd>
<kwd>feature-level fusion</kwd>
<kwd>computer-vision features</kwd>
<kwd>hierarchical age estimation</kwd>
<kwd>age group classification</kwd>
<kwd>facial aging</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>Humans have a great inherent ability to perceive visual signals around them in the environment using their eyes and beyond to analyze those signals by their brains to make some inferred decisions or actions based on their background, knowledge, and experience [<xref ref-type="bibr" rid="ref-1">1</xref>]. Nowadays, the continuing advances in computer science along with artificial intelligence technologies make it possible to imitate some human capabilities in such visual perception. Numerous techniques based on machine learning and computer vision have been proposed and elaborated aiming to successfully simulate desirable human behaviors and capabilities in receiving and processing visualized data and then making appropriate decisions or reactions.</p>
<p>In recent years, analyzing information that can be derived from human face has attracted great interest in several areas associated with pattern recognition, computer vision and image processing approaches. There is often further implicit and semantic information which can be deduced from a human face image such as gender, age, identity, race, and sentiment expressions. The problem of estimating the age of human <italic>via</italic> a face image is deemed as a very active research issue in the last decades, due to the need to provide and develop the property or ability to determine the age of individuals from their faces (see <xref ref-type="fig" rid="fig-1">Fig. 1</xref>) in many real-world applications, for example forensics, crime detection and prevention, identification of persons missing for several years, surveillance systems for suspects discovery, facial recognition and verification, access control for web content based on age, age simulation, cosmetics surgery, biometrics, and many other fields [<xref ref-type="bibr" rid="ref-2">2</xref>,<xref ref-type="bibr" rid="ref-3">3</xref>].</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Albert Einstein&#x2019;s photographs in different ages</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-1.png"/>
</fig>
<sec id="s1_1">
<label>1.1</label>
<title>Facial Aging</title>
<p>Facial age estimation can be defined as the process of assigning an age group or a specific age to a person&#x2019;s face image based on some appearance features [<xref ref-type="bibr" rid="ref-3">3</xref>]. With age progression, there may be many changes that occur in face appearance alongside aging progression. The characteristics of these changes can vary between people in their childhood and adulthood stages. During childhood stage, the most prominent changes are those which appear in the face and visibly affect its shape and size and its components, which occur because of the rapid growth of bones and skull, such as the chin becomes more salient as the cheeks extend. On the other hand, the changes after adulthood are highly associated with the soft tissue, skin texture, fine lines, wrinkles, sagging, skin elasticity loosing, and many other skin-related details, in addition to minor changes in face shape [<xref ref-type="bibr" rid="ref-4">4</xref>]. Many intrinsic and extrinsic factors may affect facial aging or appearance and make the person&#x2019;s face looks younger or older, such as genes, muscle tone, collagen ratio, sunlight, lifestyle, make-up, nutrition, cosmetic operations, weather conditions, and smoking [<xref ref-type="bibr" rid="ref-5">5</xref>]. Moreover, many research explorations have shown that facial aging may differ by population and learning facial age jointly with expression, ethnicity, and/or gender is more difficult than learning facial age disregarding these considerations [<xref ref-type="bibr" rid="ref-6">6</xref>]. All these factors and more make the age estimation process more difficult and complex. Thus, there is still an urgent need to design and develop robust algorithms integrating their strength capabilities to better confront such challenging factors and help determine the facial age as accurately as possible. <xref ref-type="fig" rid="fig-2">Figs. 2</xref> and <xref ref-type="fig" rid="fig-3">3</xref> show the usual changes occurring to young and old people with age progression.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Facial growth in children and young people with age progression [<xref ref-type="bibr" rid="ref-7">7</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-2.png"/>
</fig>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Sketch of changes occurring in the adult face with age progression [<xref ref-type="bibr" rid="ref-8">8</xref>]</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-3.png"/>
</fig>
</sec>
<sec id="s1_2">
<label>1.2</label>
<title>Traditional Computer Vision-based Age Estimation</title>
<p>Several different computer vision methods have been investigated and used for facial age estimation implicating either extracting hand-crafted features or learning deep features <italic>via</italic> Conventional Neural Network (CNN/ConvNet) architectures. The anthropometric model is focused on some measurements between the main facial components. A set of landmark points is identified on the eyes, lips, nose, eyebrows, ears, chin, and forehead. Then, a number of measurements, like axial distance, the angle between components, shortest distance, tangential distance, angle of inclination, and ratios of distances can be calculated among face components [<xref ref-type="bibr" rid="ref-9">9</xref>]. In [<xref ref-type="bibr" rid="ref-10">10</xref>,<xref ref-type="bibr" rid="ref-11">11</xref>] the authors relied on anthropometrical measurements in the process of age group classification using Artificial Neural Networks (ANNs), while in Thukral et al. [<xref ref-type="bibr" rid="ref-12">12</xref>], a hierarchical approach was applied in the process of human age estimation using 2D landmarks as shape features to train Support Vector Machine (SVM) and Support Vector Regression (SVR).</p>
<p>Active Shape Model (ASM) is a statistical model used to obtain descriptive information about face shape <italic>via</italic> a collection of facial landmark points [<xref ref-type="bibr" rid="ref-13">13</xref>]. In the age estimation process, other models are often combined with ASM in order to obtain more accurate description for a human face. The method used in [<xref ref-type="bibr" rid="ref-14">14</xref>], fuses ASM, interior angle formulation, anthropometric model, carnio-facial development, heat maps, and wrinkles together to extract aging features. Then, a CNN model is used to classify images among different age groups. Another enhanced prediction model was developed in [<xref ref-type="bibr" rid="ref-15">15</xref>] by combining both texture and shape features, and applying Partial Least-Square (PLS) for dimensionality reduction.</p>
<p>Age determination can be successfully performed using a poly regression model. Active Appearance Model (AAM) is used to extract features in many research explorations related to face images and was exploited by different researchers like Shejul et al. [<xref ref-type="bibr" rid="ref-16">16</xref>] to extract facial features for human age estimation task. It is worth noting that AAM is a statistical model distinguished from the ASM, such that it combines two different models for modeling shape and gray-level appearance. A wrinkle model was combined with AAM in [<xref ref-type="bibr" rid="ref-17">17</xref>] in order to enhance the results of basic AAM, and a CNN regressor was trained for exact age prediction. In [<xref ref-type="bibr" rid="ref-18">18</xref>], AAM, Local Binary Patterns (LBP), and Gabor Wavelets (GW) algorithms were fused together for features extraction and a novel hierarchical approach was implemented with two consecutive regressors. A skin spots feature was added in [<xref ref-type="bibr" rid="ref-19">19</xref>] using Local Phase Quantization (LPQ) algorithm. However, this feature gave a slightly negative effect on the results due to the variation of lighting among images in the used datasets.</p>
<p>AGing pattErn Subspace (AGES) was proposed by Geng et al. to be used in automatic age estimation [<xref ref-type="bibr" rid="ref-20">20</xref>]. They defined the aging pattern as a sequence of facial images of a specific person, and those images must be ordered according to the age series. Hence, rather than using independent face images, the AGES approach considered each aging pattern as one sample. Also in AGES, each image is in one specific location on the axis <italic>t</italic> which refers to the time. If a person&#x2019;s picture is available at a certain age, then it is placed in the appropriate place in the aging pattern. Otherwise, if the image is not available, its location remains empty. Each image is transformed using AAM into a vector of features, then all vectors are combined in one long vector with an empty value for each missing image. For unavailable images, the Principal Component Analysis (PCA) can be performed frequently to reconstruct the vectors of these lost images. In [<xref ref-type="bibr" rid="ref-21">21</xref>], a three-level hierarchical approach was performed using SVM and SVR with an AGES representation model, where shape, texture, and wrinkles features were fused together in one integrated model.</p>
<p>CNN is one of the commonly recently used deep learning models to analyze images, videos, and other 2D and 3D data. Images can be interpreted by CNNs as three-dimensional volumes, then, after each layer of the CNN model, the input and output volumes are represented mathematically in terms of multi-dimensional matrices. After that, the dimensions of these matrices are transformed as the image goes deeper into a CNN model. CNN structures are composed of many layers, such as convolutional layers, sub-sampling layers, and fully connected layers [<xref ref-type="bibr" rid="ref-22">22</xref>].</p>
<p>A CNN model proposed in [<xref ref-type="bibr" rid="ref-23">23</xref>] combines handcrafted features and multistage learned features of the facial images. This model includes two approaches: the first approach is based on a feature-level fusion of several local handcrafted features of wrinkles, skin with some other Biologically Inspired Features (BIFs), and the second approach is score-level fusion of feature vectors learned from a CNN with multiple layers. In [<xref ref-type="bibr" rid="ref-24">24</xref>], a new CNN architecture is introduced as Directed Acyclic Graph Convolutional Neural Networks (DAG-CNN) for estimating human age, which automatically learns discriminative features obtained from different layers of a GoogLeNet CNN [<xref ref-type="bibr" rid="ref-25">25</xref>] and VGG16 CNN [<xref ref-type="bibr" rid="ref-26">26</xref>] models and combines them together. As such, in [<xref ref-type="bibr" rid="ref-24">24</xref>], they built two variant architectures. The first is DAG-GoogLeNet based on GoogLeNet CNN and the second is DAG-VGG16 based on VGG-16. Finally, the task of estimating human age was implemented in the decision layer. In [<xref ref-type="bibr" rid="ref-27">27</xref>], a transfer learning was used to solve the problem of gender and age recognition from an image using both VGG19 and VGGFace pre-trained models. A hierarchy of deep CNNs was evaluated, which initially classifies persons by gender then predicts their age using separate male and female age prediction models.</p>
</sec>
<sec id="s1_3">
<label>1.3</label>
<title>Human Perception for Face Aging</title>
<p>With the increasing growth of real-world applications in the last decade, there has been also a growing interest in studying visual perception of human in facial age estimation tasks, and comparing the ability of machines versus humans in such tasks. Some researchers, such as [<xref ref-type="bibr" rid="ref-28">28</xref>], explored the ability of humans to estimate the age of subjects from their face images. Such that they showed a set of pictures to a group of participants and asked them to guess the age of each person shown in a picture based on their facial appearance. Simultaneously, they used the same images in building a separate machine learning model. After conducting performance evaluation and comparison, they concluded that the performance of human perception and machine learning attained close results in the age prediction task. Han et al. [<xref ref-type="bibr" rid="ref-29">29</xref>] proposed a framework for automatically estimating demographics from human face based on biologically inspired features (BIF). Furthermore, they used crowdsourcing mechanism to estimate human perceptual performance in demographic prediction on a variety of aging datasets, including FG-NET, FERET, MORPH II, PCSO, and LFW. This enables comparison between computer and human abilities in predicting three demographics (age, gender, and race). They found that their proposed framework can closely match the performance of human in demographic prediction in all three demographic prediction tasks and it performs slightly better than human with PCSO and MORPH II datasets.</p>
<p>In [<xref ref-type="bibr" rid="ref-30">30</xref>], a study was conducted on the use of a CNN-based model for estimating women&#x0027;s face ages automatically, and further for comparing machine and human performance in terms of which face regions they are much focused on during performing the same task on the same image. The reported human prediction was almost as accurate as the machine prediction using VGG16 model (i.e., 60% accuracy for the CNN model against 61% for human prediction). Their experimental results of tracking human&#x0027;s eyes showed that, when a participant focused their gaze more on the eyes or the mouth regions, their accuracy increased in estimating a person&#x0027;s age; conversely, their accuracy degraded when their gaze concentrated more on other facial skin regions. This may provide some clues to the fact that, indeed, humans may be able to accurately estimate age of others based on some semantic facial features. It also indicated that human-vision capabilities are still competitive and cannot be neglected, even when compared with nowadays sophisticated computer-vision capabilities, and, thus, they can be rather utilized to be integrated with computer-vision capabilities or perhaps to be emulated for achieving augmented performance in such a difficult age estimation task. As such, a reasonable open question can be: What would be the performance of age estimation in the fusion of the capabilities of both humans and machines?</p>
<p>Although some researchers in the age estimation field have paid attention to studying human perception and interpretation of facial aging and comparing their abilities with machines, exploiting humans&#x2019; vision capabilities in analyzing face images and deriving detailed semantic features for age estimation purposes is yet to be investigated, which may, in turn, help bridging the semantic gap and effectively improve the latest machine-based performance. However, several previous biometrics research studies concerned with recognizing human face [<xref ref-type="bibr" rid="ref-31">31</xref>], identifying subjects [<xref ref-type="bibr" rid="ref-32">32</xref>] and recognizing gait signature [<xref ref-type="bibr" rid="ref-33">33</xref>] have employed human-vision in analyzing and annotating images or videos to derive semantic biometric traits. Their results proved the powerful and effectiveness of such semantic traits in improving model-based performance.</p>
<p><bold>The main contributions</bold> of this paper can be summarized as follows:<list list-type="bullet"><list-item>
<p>Exploiting human-vision capabilities in the process of extracting a novel set of semantic features to be used for a proposed semantic-based age estimation approach in fusion with different combinations of traditional vision-based features.</p></list-item><list-item>
<p>Analyzing our proposed semantic features using different statistical tests to investigate and select the most significant set of semantic features for high-level (age group classification) and low-level (regression for exact age prediction) stages.</p></list-item><list-item>
<p>Utilizing these semantic features in augmenting and improving the performance of traditional computer vision models in predicting exact age and age groups/levels via face image analysis, allowing for exploring and proving that these two forms of facial features (i.e., traditional vision-based and semantic human-based) are differently informative and effectively integrative in human age estimation tasks.</p></list-item><list-item>
<p>Providing an initial research study that may establish further promising research tracks to be undertaken by researchers in the future utilizing our obtained results and findings in how to supplement computer vision-based features with human-based features for augmented facial-based human age estimation, in a way, filling/bridging the semantic gap between humans and machines.</p></list-item></list></p>
<p>The remainder of this paper is organized as follows. Section 2 explains images dataset, proposed novel semantic features, computer vision-based features, hierarchical multi-level age estimation and evaluation metrics used to evaluate proposed approaches. Experiments and results are presented in Section 3. Section 4 provides a conclusion and future work.</p>
</sec>
</sec>
<sec id="s2">
<label>2</label>
<title>Methods and Materials</title>
<p>In this section, we will explain the face images dataset, computer-vision features, semantic features, hierarchical multi-level estimation approach, and the metrics used for model evaluation.</p>
<sec id="s2_1">
<label>2.1</label>
<title>Dataset</title>
<p>To evaluate our proposed approach, we used a ready-made and publicly available standard dataset called Face and Gesture Recognition Network (FG-NET) [<xref ref-type="bibr" rid="ref-34">34</xref>]. It is one of the most popular and frequently used datasets in the field of facial aging and age estimation [<xref ref-type="bibr" rid="ref-3">3</xref>,<xref ref-type="bibr" rid="ref-35">35</xref>,<xref ref-type="bibr" rid="ref-36">36</xref>]. It contains 1002 face images for 82 individuals, 48 males and 34 females. Individuals in FG-NET dataset have from 6 to 18 color or grayscale images at different ages of the same individuals, with an average of 12 images per individual. The range of ages in FG-NET dataset varies between 0 and 69 years old. There are 68 landmark points provided with each image in FG-NET that represent and localize the face boundaries, eyebrows, lips, eyes, and nose.</p>
<p>Most FG-NET images are collected by scanning the personal photos of individuals from multiple races. Therefore, the quality of images depends on the skills of photographers, camera used, lighting when taking photos, and the quality of photographic paper. This consequently causes a variation among images in quality, resolution, expression, pose, viewpoint, and illumination. In addition, some occlusions appear in a number of images, like makeup, spectacles, hats, beards, and mustaches that may cover or blur the aging signs in some areas of skin. All these conditions posed additional challenges in building an adaptable model capable of analyzing facial aging and achieving desired results. A sample of images for one individual in FG-NET at variant ages is shown in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Sample of images for a single person in FG-NET dataset, at ages 2, 5, 8, 10, 14, 16, 18, 19, 22, 28, 29, 33, 40, 43, and 43 years old, shown from left to right, respectively</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-4.png"/>
</fig>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Extracting Semantic Features</title>
<p>In this research, the Figure-Eight crowdsourcing platform [<xref ref-type="bibr" rid="ref-37">37</xref>] was used to collect our novel semantic features. Whilst, most of these semantic aging features were inspired by different forensic concepts of face aging [<xref ref-type="bibr" rid="ref-5">5</xref>], we designed web-based annotation forms and used them to show each FG-NET image to several annotators on the platform and asked them to annotate the displayed image by choosing the most applicable label from a set of given descriptive labels for each facial feature, in such a way describing the degree of presence of each semantic feature on the displayed subject face. These labels are either ordinal such as (None, Minimal, Fair, Marked, or Prominent) and (No wrinkles, Wrinkles with expressions, Wrinkles with rest, or Prominent wrinkles) or binary nominal labels such as (male or female) and (absent or present). The annotator should choose only one most appropriate option/label per feature, where each face image was labeled by multiple annotators. Subsequently, the labels were rescaled using z-score standardization given as:</p>
<p><disp-formula id="eqn-1">
<label>(1)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-1.png"/><tex-math id="tex-eqn-1"><![CDATA[Z = \displaystyle{{\left( {x - \mu } \right)} \over {\left( \sigma \right)}}]]></tex-math>--><mml:math id="mml-eqn-1" display="block"><mml:mi>Z</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>&#x03BC;</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03C3;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-1">
<!--<alternatives><inline-graphic xlink:href="ieqn-1.tif"/><tex-math id="tex-ieqn-1"><![CDATA[x]]></tex-math>--><mml:math id="mml-ieqn-1"><mml:mi>x</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is a feature vector, <inline-formula id="ieqn-2"><!--<alternatives><inline-graphic xlink:href="ieqn-2.tif"/><tex-math id="tex-ieqn-2"><![CDATA[\mu]]></tex-math>--><mml:math id="mml-ieqn-2"><mml:mi>&#x03BC;</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is the mean and <inline-formula id="ieqn-3">
<!--<alternatives><inline-graphic xlink:href="ieqn-3.tif"/><tex-math id="tex-ieqn-3"><![CDATA[\sigma]]></tex-math>--><mml:math id="mml-ieqn-3"><mml:mi>&#x03C3;</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is the standard deviation. Then average-labels were calculated for every feature per face image. To facilitate the annotation task, we provided a few instructions with a visual example of an annotated image and grouped related features according to each targeted facial part. The designed annotation task form is displayed in <xref ref-type="fig" rid="fig-5">Fig. 5</xref>.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Designed annotation form for semantic feature acquisition</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-5.png"/>
</fig>
<p>This study exploited the human-vision capabilities of a set of people to analyze and extract changes occurring in texture and thickness of facial soft tissue during age progression and investigated their effects on human facial morphology. Moreover, global aging signs like beard, mustache, gray hair, and pallor were included, in addition to other aging-related features introduced in [<xref ref-type="bibr" rid="ref-38">38</xref>&#x2013;<xref ref-type="bibr" rid="ref-40">40</xref>]. <xref ref-type="table" rid="table-1">Tab. 1</xref> shows all proposed 32 semantic face features with their corresponding labels.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Proposed semantic features with their labels grouped according to facial components</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Face<break/>Component</th>
<th>Semantic Feature Description</th>
<th>Abbreviation</th>
<th>Semantic Labels</th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="2">Forehead</td>
<td>Lines and wrinkles forming horizontally on forehead</td>
<td>F1</td>
<td rowspan="17">None, Minimal, Fair, Marked, or Prominent</td>
</tr>
<tr>
<td>Elongation or extension of forehead height</td>
<td>F2</td>
</tr>
<tr>
<td rowspan="2">Eyebrows</td>
<td>Vertical or slightly diagonal lines between eyebrows</td>
<td>F3</td>
</tr>
<tr>
<td>Sagging or dropping of eyebrows</td>
<td>F4</td>
</tr>
<tr>
<td rowspan="6">Eyes</td>
<td>Lines and wrinkles around outer corners of eyes</td>
<td>F5</td>
</tr>
<tr>
<td>Eye bags above and below eyes</td>
<td>F6</td>
</tr>
<tr>
<td>Sagging or dropping of upper eyelids</td>
<td>F7</td>
</tr>
<tr>
<td>Sunken or hollowing eyes</td>
<td>F8</td>
</tr>
<tr>
<td>Sunken or hollowing lower eyelids</td>
<td>F9</td>
</tr>
<tr>
<td>Dark circles under eyes</td>
<td>F10</td>
</tr>
<tr>
<td rowspan="2">Nose</td>
<td>Bulbous nose</td>
<td>F11</td>
</tr>
<tr>
<td>Lines and wrinkles on nose</td>
<td>F12</td>
</tr>
<tr>
<td rowspan="6">Mouth</td>
<td>Vertical lines above and below lips</td>
<td>F13</td>
</tr>
<tr>
<td>Elongation or expansion of region between nose and mouth</td>
<td>F14</td>
</tr>
<tr>
<td>Smile lines or nasolabial folds</td>
<td>F15</td>
</tr>
<tr>
<td>Curved lines around mouth corners</td>
<td>F16</td>
</tr>
<tr>
<td>Thinning and flattening of lips</td>
<td>F17</td>
</tr>
<tr>
<td>Mustache</td>
<td>F18</td>
<td>No (absent) or Yes (present)</td>
</tr>
<tr>
<td rowspan="3">Cheeks</td>
<td>Sagging or dropping of cheeks</td>
<td>F19</td>
<td rowspan="7">None, Minimal, Fair, Marked, or Prominent</td>
</tr>
<tr>
<td>Sunken or hollowing of cheeks</td>
<td>F20</td>
</tr>
<tr>
<td>Lines and wrinkles on cheeks</td>
<td>F21</td>
</tr>
<tr>
<td>Jawlines</td>
<td>Jowls (small focal accumulation of fat in the lower cheek overlying jawbone)</td>
<td>F22</td>
</tr>
<tr>
<td rowspan="4">Chin</td>
<td>Sagging or dropping of chin</td>
<td>F23</td>
</tr>
<tr>
<td>Marionette lines</td>
<td>F24</td>
</tr>
<tr>
<td>Lines and wrinkles on chin</td>
<td>F25</td>
</tr>
<tr>
<td>Beard</td>
<td>F26</td>
<td>No (absent) or Yes (present)</td>
</tr>
<tr>
<td rowspan="6">Whole face</td>
<td>Elasticity losing</td>
<td>F27</td>
<td rowspan="4">None, Minimal, Fair, Marked, or Prominent</td>
</tr>
<tr>
<td>Age spots or liver spots</td>
<td>F28</td>
</tr>
<tr>
<td>Pallor of skin, lips and eyes color</td>
<td>F29</td>
</tr>
<tr>
<td>Gray hair in eyebrows and eyelashes (mustache and beard if any)</td>
<td>F30</td>
</tr>
<tr>
<td>Wrinkling degree of entire face</td>
<td>F31</td>
<td>No Wrinkles, Wrinkles with Expressions, Wrinkles with rest, or Prominent Wrinkles</td>
</tr>
<tr>
<td>Gender</td>
<td>F32</td>
<td>Male or Female</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Quality assurance services offered by the Figure-Eight platform were all set up to ensure data reliability, such as limiting tasks only for the highest accuracy and most experienced annotators, limiting the maximum number of judgments available for each annotator, pre-test questions for checking annotator understanding and excluding ineligible annotators, setting up detection and prevention of the extra-fast or random responses, and deploying our task for geographically unconstrained annotators to enable better reflection of average human-annotator perception. <xref ref-type="table" rid="table-2">Tab. 2</xref> provides a summary of collected semantic data.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Summary of collected semantic data</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Semantic Data Summary</th>
<th>Quantity/Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Annotations per image</td>
<td>5</td>
</tr>
<tr>
<td>Total annotations for all images</td>
<td>5010</td>
</tr>
<tr>
<td>Total labels (32 labels per annotation)</td>
<td>160320</td>
</tr>
<tr>
<td>Maximum annotations allowed per annotator</td>
<td>5</td>
</tr>
<tr>
<td>Total number of annotators</td>
<td>1259</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To investigate and select the most significant semantic features in classification and regression stages, we applied two different statistical analysis tests:</p>
<sec id="s2_2_1">
<label>2.2.1</label>
<title>ANalysis Of VAriance (ANOVA)</title>
<p>Analysis of variance (ANOVA) is a statistical test used to check whether the means of at least two groups differ from each other or not [<xref ref-type="bibr" rid="ref-41">41</xref>]. We performed ANOVA on our semantic features to measure the significance of each in distinguishing between different age groups in classification stage with a 0.01 significance level.</p>
<p>The <italic>p</italic>-value for all semantic features is much less than the significance level, 0.01. Therefore, we rejected the null hypothesis for all features and accepted the alternative hypothesis states that there is a difference in the means of at least two age groups. <xref ref-type="table" rid="table-3">Tab. 3</xref> introduces selected semantic features in descending order according to their effectiveness by F-ratio and corresponding <italic>p</italic>-value.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Semantic features arranged descendingly by F-ratio and P-value according to their effectiveness</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Semantic<break/>Feature</th>
<th>F1</th>
<th>F27</th>
<th>F5</th>
<th>F31</th>
<th>F2</th>
<th>F4</th>
<th>F7</th>
<th>F9</th>
<th>F8</th>
<th>F25</th>
<th>F6</th>
<th>F24</th>
<th>F19</th>
<th>F3</th>
<th>F21</th>
<th>F22</th>
</tr>
</thead>
<tbody>
<tr>
<td>P-value</td>
<td>1<break/>e-169</td>
<td>1.20<break/>e-168</td>
<td>6.59<break/>e-161</td>
<td>1.30<break/>e-152</td>
<td>1.66<break/>e-149</td>
<td>2.94<break/>e-149</td>
<td>7.15<break/>e-149</td>
<td>1.23<break/>e-144</td>
<td>4.49<break/>e-143</td>
<td>3.95<break/>e-142</td>
<td>1.31<break/>e-139</td>
<td>4.01<break/>e-136</td>
<td>3.93<break/>e-134</td>
<td>1.90<break/>e-133</td>
<td>2.89<break/>e-130</td>
<td>2.27<break/>e-129</td>
</tr><tr>
<td>F-ratio</td>
<td>397.2</td>
<td>393.6</td>
<td>368.1</td>
<td>341.7</td>
<td>332.1</td>
<td>331.3</td>
<td>330.2</td>
<td>317.3</td>
<td>312.6</td>
<td>309.8</td>
<td>302.4</td>
<td>292.3</td>
<td>286.5</td>
<td>284.6</td>
<td>275.6</td>
<td>273.1</td>
</tr><tr>
<td>Semantic<break/>Feature</td>
<td>F23</td>
<td>F13</td>
<td>F16</td>
<td>F20</td>
<td>F30</td>
<td>F10</td>
<td>F17</td>
<td>F15</td>
<td>F29</td>
<td>F28</td>
<td>F12</td>
<td>F14</td>
<td>F11</td>
<td>F18</td>
<td>F26</td>
<td>F32</td>
</tr>
<tr>
<td>P-value</td>
<td>1.41<break/>e-128</td>
<td>6.46<break/>e-125</td>
<td>7.61<break/>e-125</td>
<td>8.93<break/>e-125</td>
<td>1.69<break/>e-120</td>
<td>8.54<break/>e-114</td>
<td>8.73<break/>e-114</td>
<td>3.34<break/>e-108</td>
<td>2.21<break/>e-107</td>
<td>8.16<break/>e-103</td>
<td>9.81<break/>e-101</td>
<td>6.74<break/>e-99</td>
<td>1.36<break/>e-94</td>
<td>8.64<break/>e-28</td>
<td>2.67<break/>e-09</td>
<td>2.24<break/>e-06</td>
</tr>
<tr>
<td>F-ratio</td>
<td>270.8</td>
<td>260.7</td>
<td>260.5</td>
<td>260.3</td>
<td>248.7</td>
<td>231</td>
<td>230.97</td>
<td>216.6</td>
<td>214.5</td>
<td>203.1</td>
<td>197.98</td>
<td>193.5</td>
<td>183.1</td>
<td>45.9</td>
<td>14.58</td>
<td>9.8</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2_2_2">
<label>2.2.2</label>
<title>Pearson Correlation Coefficient</title>
<p>Pearson&#x2019;s correlation coefficient is used in statistics to measures strength and direction of linear correlation between two continuous variables <inline-formula id="ieqn-4">
<!--<alternatives><inline-graphic xlink:href="ieqn-4.tif"/><tex-math id="tex-ieqn-4"><![CDATA[x]]></tex-math>--><mml:math id="mml-ieqn-4"><mml:mi>x</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> and <inline-formula id="ieqn-5">
<!--<alternatives><inline-graphic xlink:href="ieqn-5.tif"/><tex-math id="tex-ieqn-5"><![CDATA[y]]></tex-math>--><mml:math id="mml-ieqn-5"><mml:mi>y</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> [<xref ref-type="bibr" rid="ref-42">42</xref>,<xref ref-type="bibr" rid="ref-43">43</xref>]. Assuming <inline-formula id="ieqn-6">
<!--<alternatives><inline-graphic xlink:href="ieqn-6.tif"/><tex-math id="tex-ieqn-6"><![CDATA[\bar x]]></tex-math>--><mml:math id="mml-ieqn-6"><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> and <inline-formula id="ieqn-7">
<!--<alternatives><inline-graphic xlink:href="ieqn-7.tif"/><tex-math id="tex-ieqn-7"><![CDATA[\bar y]]></tex-math>--><mml:math id="mml-ieqn-7"><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> are the means of first and second variables, respectively, n is the number of pairs in <inline-formula id="ieqn-8">
<!--<alternatives><inline-graphic xlink:href="ieqn-8.tif"/><tex-math id="tex-ieqn-8"><![CDATA[x]]></tex-math>--><mml:math id="mml-ieqn-8"><mml:mi>x</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> and <inline-formula id="ieqn-9">
<!--<alternatives><inline-graphic xlink:href="ieqn-9.tif"/><tex-math id="tex-ieqn-9"><![CDATA[y]]></tex-math>--><mml:math id="mml-ieqn-9"><mml:mi>y</mml:mi></mml:math>
<!--</alternatives>--></inline-formula>, the formula of Pearson&#x0027;s correlation coefficient <inline-formula id="ieqn-10">
<!--<alternatives><inline-graphic xlink:href="ieqn-10.tif"/><tex-math id="tex-ieqn-10"><![CDATA[r]]></tex-math>--><mml:math id="mml-ieqn-10"><mml:mi>r</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is defined as in <xref ref-type="disp-formula" rid="eqn-2">Eq. (2)</xref>. When the correlation is closer to &#x002B;1, this indicates a stronger positive correlation and highly correlated variables, whereas when it is closer to -1, this means a stronger but negative correlation between variables. Correspondingly, the correlation becomes either way weaker as it closes to zero.</p>
<p><disp-formula id="eqn-2">
<label>(2)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-2.png"/><tex-math id="tex-eqn-2"><![CDATA[r = \displaystyle{{\sum\limits_{i = 1}^n {} \left( {{x_i} - \bar x} \right)\left( {{y_i} - \bar y} \right)} \over {\sqrt {\sum\limits_{i = 1}^n {{{\left( {{x_i} - \bar x} \right)}^2}\sum\limits_{i = 1}^n {{{\left( {{y_i} - \bar y} \right)}^2}} } } }}]]></tex-math>--><mml:math id="mml-eqn-2" display="block"><mml:mi>r</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mrow><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:msqrt><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mrow></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>As shown in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>, the correlation between age and each of our semantic features is strongly positive, except for three features, which are gender and presence of mustache or beard. This is due to that very few subjects appear in FG-NET dataset images with mustache or beard, such that only 36 images show subjects with a mustache and 31 with a beard. Furthermore, there is a lack of a direct or sensible relationship between human gender and age, due to the natural stability of gender over the lifetime. Nevertheless, extended analysis and exploration of different aging characteristics with respect to gender could enable more understanding of the similarities and differences between males and females (or average male and average female) in aging progression changes or signs from different perspectives and various aspects, such as what, when, where, and how similar or different those aging characteristics are expected to be, which could, in turn, be considered to help age estimation. Thus, these three semantic features were excluded from regression stage.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>Correlation matrix between semantic features and age and between each other using Pearson&#x0027;s r correlation coefficient</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-6.png"/>
</fig>
<p><xref ref-type="table" rid="table-4">Tab. 4</xref> provides semantic features in descending order based on their correlation strength with age in addition to corresponding p-value for each feature at 0.01 confidence level.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Semantic features arranged in descending order according to their correlation strength with age</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Semantic<break/>Feature</th>
<th>F27</th>
<th>F5</th>
<th>F1</th>
<th>F31</th>
<th>F7</th>
<th>F6</th>
<th>F4</th>
<th>F25</th>
<th>F8</th>
<th>F9</th>
<th>F24</th>
<th>F2</th>
<th>F19</th>
<th>F16</th>
<th>F22</th>
<th>F21</th>
</tr>
</thead>
<tbody>
<tr>
<td>P-value</td>
<td>2.34<break/>e-194</td>
<td>4.01<break/>e-184</td>
<td>1.56<break/>e-178</td>
<td>3.73<break/>e-175</td>
<td>2.60<break/>e-173</td>
<td>1.30<break/>e-166</td>
<td>9.67<break/>e-161</td>
<td>1.60<break/>e-160</td>
<td>1.95<break/>e-160</td>
<td>4<break/>e-160</td>
<td>2.24<break/>e-157</td>
<td>2.88<break/>e-156</td>
<td>6.24<break/>e-152</td>
<td>1.08<break/>e-150</td>
<td>1.33<break/>e-150</td>
<td>3.02<break/>e-150</td>
</tr><tr>
<td>Pearson&#x2019;s r</td>
<td>0.77</td>
<td>0.75</td>
<td>0.75</td>
<td>0.74</td>
<td>0.74</td>
<td>0.73</td>
<td>0.72</td>
<td>0.72</td>
<td>0.72</td>
<td>0.72</td>
<td>0.71</td>
<td>0.71</td>
<td>0.71</td>
<td>0.7</td>
<td>0.7</td>
<td>0.7</td>
</tr><tr>
<td>Semantic<break/>Feature</td>
<td>F23</td>
<td>F13</td>
<td>F20</td>
<td>F3</td>
<td>F10</td>
<td>F15</td>
<td>F17</td>
<td>F29</td>
<td>F28</td>
<td>F12</td>
<td>F14</td>
<td>F11</td>
<td>F30</td>
<td>F18</td>
<td>F26</td>
<td>F32</td>
</tr>
<tr>
<td>P-value</td>
<td>3.79<break/>e-149</td>
<td>5.10<break/>e-145</td>
<td>2.12<break/>e-144</td>
<td>3.40<break/>e-140</td>
<td>3.21<break/>e-136</td>
<td>6.11<break/>e-136</td>
<td>1.56<break/>e-132</td>
<td>1.06<break/>e-126</td>
<td>4.66<break/>e-122</td>
<td>7.20<break/>e-116</td>
<td>5.96<break/>e-115</td>
<td>1.54<break/>e-110</td>
<td>1.10<break/>e-108</td>
<td>5.02<break/>e-29</td>
<td>5.37<break/>e-07</td>
<td>0.656<break/>e-3</td>
</tr>
<tr>
<td>Pearson&#x2019;s r</td>
<td>0.7</td>
<td>0.69</td>
<td>0.69</td>
<td>0.69</td>
<td>0.68</td>
<td>0.68</td>
<td>0.67</td>
<td>0.66</td>
<td>0.65</td>
<td>0.64</td>
<td>0.64</td>
<td>0.63</td>
<td>0.62</td>
<td>0.34</td>
<td>0.16</td>
<td>0.11</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Extracting Computer-vision Features</title>
<p>To prepare the data images for computer vision-based feature extraction stage, all colored images were converted into grayscale. Next, since the FG-NET contains images from personal image collections of subjects, pose variation is likely observed in many subject images. Thus, images were rotated to a standard face pose. Image rotation was conducted based on the coordinates of the center of the two eyes to align all images vertically, where the rotation angle <inline-formula id="ieqn-11">
<!--<alternatives><inline-graphic xlink:href="ieqn-11.tif"/><tex-math id="tex-ieqn-11"><![CDATA[\theta]]></tex-math>--><mml:math id="mml-ieqn-11"><mml:mi>&#x03B8;</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is computed as follows:</p>
<p><disp-formula id="eqn-3">
<label>(3)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-3.png"/><tex-math id="tex-eqn-3"><![CDATA[\theta = {\tan ^{ - 1}}\displaystyle{{{Y_L} - {Y_R}} \over {{X_L} - {X_R}}}]]></tex-math>--><mml:math id="mml-eqn-3" display="block"><mml:mi>&#x03B8;</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mrow><mml:msup><mml:mi>tan</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mi>L</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi>X</mml:mi><mml:mi>L</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>X</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-12">
<!--<alternatives><inline-graphic xlink:href="ieqn-12.tif"/><tex-math id="tex-ieqn-12"><![CDATA[\left( {{X_R},{Y_R}} \right)]]></tex-math>--><mml:math id="mml-ieqn-12"><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>X</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> and <inline-formula id="ieqn-13">
<!--<alternatives><inline-graphic xlink:href="ieqn-13.tif"/><tex-math id="tex-ieqn-13"><![CDATA[\left( {{X_L},{Y_L}} \right)]]></tex-math>--><mml:math id="mml-ieqn-13"><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>X</mml:mi><mml:mi>L</mml:mi></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mi>L</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> are the position of the right and left eyes, respectively. <xref ref-type="fig" rid="fig-7">Fig. 7</xref> provides an example of a face image before and after preprocessing.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>(a) is the original image showing eyes coordinates in yellow color. (b) is the same image after rotation and grayscale conversion</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-7.png"/>
</fig>
<sec id="s2_3_1">
<label>2.3.1</label>
<title>Active Appearance Model</title>
<p>Active Appearance Model (AAM) is a computer-vision model proposed by Cootes et al. [<xref ref-type="bibr" rid="ref-44">44</xref>]. It combines shape and gray-level appearance statistical models to represent and interpret objects in images. It was developed and used in many applications such as objects tracking, gait analysis, human eyes modeling, facial expression recognition, and medical image segmentation [<xref ref-type="bibr" rid="ref-45">45</xref>]. We have employed AAM [<xref ref-type="bibr" rid="ref-46">46</xref>] proposed for fast AAM fitting in-the-wild.</p>
<p>In this work, to build the shape model, a set of 68 landmark points (<italic>X</italic><sub>1</sub>, <italic>Y</italic><sub>1</sub>, <italic>X</italic><sub>2</sub>, <italic>Y</italic><sub>2</sub>, &#x2026;., <italic>X</italic><sub>68</sub>, <italic>Y</italic><sub>68</sub>) describing the shape of a human face is required across <italic>D</italic> training images. Procrustes analysis was applied to remove similarity transformation from original shapes and get <italic>D</italic> similarity-free shapes. Then, PCA was performed on these shapes to obtain the shape model <inline-formula id="ieqn-14">
<!--<alternatives><inline-graphic xlink:href="ieqn-14.tif"/><tex-math id="tex-ieqn-14"><![CDATA[\hat s]]></tex-math>--><mml:math id="mml-ieqn-14"><mml:mrow><mml:mover><mml:mi>s</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula>, defined by a set of shape eigenvectors <inline-formula id="ieqn-15">
<!--<alternatives><inline-graphic xlink:href="ieqn-15.tif"/><tex-math id="tex-ieqn-15"><![CDATA[S]]></tex-math>--><mml:math id="mml-ieqn-15"><mml:mi>S</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> and the mean shape <inline-formula id="ieqn-16">
<!--<alternatives><inline-graphic xlink:href="ieqn-16.tif"/><tex-math id="tex-ieqn-16"><![CDATA[{s_0}]]></tex-math>--><mml:math id="mml-ieqn-16"><mml:mrow><mml:msub><mml:mi>s</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula>. The statistical shape model can be expressed as:</p>
<p><disp-formula id="eqn-4">
<label>(4)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-4.png"/><tex-math id="tex-eqn-4"><![CDATA[$\hat s = {s_0} + SP,\,{\rm where} \, P = {S^T}\left( {s - {s_0}} \right)$]]></tex-math>--><mml:math id="mml-eqn-4" display="block"><mml:mrow><mml:mover accent='true'><mml:mi>s</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:msub><mml:mi>s</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mi>S</mml:mi><mml:mi>P</mml:mi><mml:mo>,</mml:mo><mml:mtext>where</mml:mtext><mml:mi>P</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mi>S</mml:mi><mml:mi>T</mml:mi></mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>s</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:math> 
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-17">
<!--<alternatives><inline-graphic xlink:href="ieqn-17.tif"/><tex-math id="tex-ieqn-17"><![CDATA[s]]></tex-math>--><mml:math id="mml-ieqn-17"><mml:mi>s</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is a new similarity-free shape and <inline-formula id="ieqn-18">
<!--<alternatives><inline-graphic xlink:href="ieqn-18.tif"/><tex-math id="tex-ieqn-18"><![CDATA[p]]></tex-math>--><mml:math id="mml-ieqn-18"><mml:mi>p</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> contains the shape parameters of a deformable model.</p>
<p>Building the appearance model requires removing variation of shapes from the texture. This was achieved by warping each texture <italic>I</italic> to the reference frame obtained using the mean shape. After that, PCA was performed on shape-free textures to obtain the appearance model <inline-formula id="ieqn-19">
<!--<alternatives><inline-graphic xlink:href="ieqn-19.tif"/><tex-math id="tex-ieqn-19"><![CDATA[\hat I]]></tex-math>--><mml:math id="mml-ieqn-19"><mml:mrow><mml:mover><mml:mi>I</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow></mml:math><!--</alternatives>--></inline-formula> defined by a set of appearance eigenvectors <inline-formula id="ieqn-20">
<!--<alternatives><inline-graphic xlink:href="ieqn-20.tif"/><tex-math id="tex-ieqn-20"><![CDATA[A]]></tex-math>--><mml:math id="mml-ieqn-20"><mml:mi>A</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> and the mean normalized appearance vector <inline-formula id="ieqn-21">
<!--<alternatives><inline-graphic xlink:href="ieqn-21.tif"/><tex-math id="tex-ieqn-21"><![CDATA[{A_0}]]></tex-math>--><mml:math id="mml-ieqn-21"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula>. By assuming <inline-formula id="ieqn-22">
<!--<alternatives><inline-graphic xlink:href="ieqn-22.tif"/><tex-math id="tex-ieqn-22"><![CDATA[c]]></tex-math>--><mml:math id="mml-ieqn-22"><mml:mi>c</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> represents the appearance parameters, the statistical appearance model can defined as:</p>
<p><disp-formula id="eqn-5">
<label>(5)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-5.png"/><tex-math id="tex-eqn-5"><![CDATA[\hat I = {A_0} + Ac,\,{\rm where}\,c = {A^T}\left( {I - {A_0}} \right)]]></tex-math>--><mml:math id="mml-eqn-5" display="block"><mml:mrow><mml:mover><mml:mi>I</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mo>&#x003D;</mml:mo><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mo>&#x002B;</mml:mo><mml:mi>A</mml:mi><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mspace width="thinmathspace"></mml:mspace><mml:mrow><mml:mi mathvariant="normal">w</mml:mi><mml:mi mathvariant="normal">h</mml:mi><mml:mi mathvariant="normal">e</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:mi mathvariant="normal">e</mml:mi></mml:mrow><mml:mspace width="thinmathspace"></mml:mspace><mml:mi>c</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mrow><mml:msup><mml:mi>A</mml:mi><mml:mi>T</mml:mi></mml:msup></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>Fast-Forward algorithm was used to fit the AAM model to extract global facial shape and appearance from each image. <xref ref-type="fig" rid="fig-8">Fig. 8</xref> below illustrates detected landmark points with corresponding appearance.</p>
<fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>(a) landmark points define facial shape, (b) Shape-free patch represents face appearance</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-8.png"/>
</fig>
</sec>
<sec id="s2_3_2">
<label>2.3.2</label>
<title>Geometric Ratios</title>
<p>Craniofacial changes occur in human face with age progression and convey much information about facial aging. They are useful in distinguishing children from adults and studying facial aging signs related to children, who are characterized by their rapid geometric growth [<xref ref-type="bibr" rid="ref-47">47</xref>].</p>
<p>Due to the fact that the facial shape may be affected by expressions and pose, this might influence on the localization of facial landmark points. Therefore, the ratios between distances were used to measure human craniofacial growth rather than solely using distances. We calculated 19 distances D <italic>via</italic> 23 landmark points and used them to form 18 different ratios R as geometric features as in [<xref ref-type="bibr" rid="ref-47">47</xref>,<xref ref-type="bibr" rid="ref-48">48</xref>]. All landmark points in this section were produced using AAM as demonstrated in the previous section, except for point number 23, which was manually localized due to hair covering the forehead or bald appearing for some individuals in the FG-NET dataset. The ratio R between two distances is calculated as:</p>
<p><disp-formula id="eqn-6">
<label>(6)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-6.png"/><tex-math id="tex-eqn-6"><![CDATA[R = \displaystyle{{{D_1}} \over {{D_2}}},\,{\rm where }\,{D_i}\left( {p,q} \right) = \sqrt {{{\left( {{p_1} - {q_1}} \right)}^2} + {{\left( {{p_2} - {q_2}} \right)}^2}}]]></tex-math>--><mml:math id="mml-eqn-6" display="block"><mml:mi>R</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:mo>,</mml:mo><mml:mspace width="thinmathspace"></mml:mspace><mml:mrow><mml:mi mathvariant="normal">w</mml:mi><mml:mi mathvariant="normal">h</mml:mi><mml:mi mathvariant="normal">e</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:mi mathvariant="normal">e</mml:mi></mml:mrow><mml:mspace width="thinmathspace"></mml:mspace><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>q</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003D;</mml:mo><mml:msqrt><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>q</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mo>&#x002B;</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>q</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:msqrt></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-23">
<!--<alternatives><inline-graphic xlink:href="ieqn-23.tif"/><tex-math id="tex-ieqn-23"><![CDATA[{p_1}]]></tex-math>--><mml:math id="mml-ieqn-23"><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula>, <inline-formula id="ieqn-24">
<!--<alternatives><inline-graphic xlink:href="ieqn-24.tif"/><tex-math id="tex-ieqn-24"><![CDATA[{p_2}]]></tex-math>--><mml:math id="mml-ieqn-24"><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula>, <inline-formula id="ieqn-25">
<!--<alternatives><inline-graphic xlink:href="ieqn-25.tif"/><tex-math id="tex-ieqn-25"><![CDATA[{q_1}]]></tex-math>--><mml:math id="mml-ieqn-25"><mml:mrow><mml:msub><mml:mi>q</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> and <inline-formula id="ieqn-26">
<!--<alternatives><inline-graphic xlink:href="ieqn-26.tif"/><tex-math id="tex-ieqn-26"><![CDATA[{q_2}]]></tex-math>--><mml:math id="mml-ieqn-26"><mml:mrow><mml:msub><mml:mi>q</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> are coordinates of p and q landmark points. <xref ref-type="fig" rid="fig-9">Fig. 9</xref> presents 23 landmark points used to calculate the ratios of geometric features and <xref ref-type="fig" rid="fig-10">Fig. 10</xref> illustrates 18 computed shape ratios of distances between different landmark points. The difference between an adult and a child in <italic>R</italic><sub>6</sub> ratio between face width and height eyebrow center and chin points is shown in <xref ref-type="fig" rid="fig-11">Fig. 11</xref>.</p>
<fig id="fig-9">
<label>Figure 9</label>
<caption>
<title>23 landmark points for calculating ratios between distances</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-9.png"/>
</fig>
<fig id="fig-10">
<label>Figure 10</label>
<caption>
<title>18 Ratios between different landmark pair points distances</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-10.png"/>
</fig>
<fig id="fig-11">
<label>Figure 11</label>
<caption>
<title>The difference between an adult and a child using R6 ratio</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-11.png"/>
</fig>
</sec>
<sec id="s2_3_3">
<label>2.3.3</label>
<title>Local Binary Patterns</title>
<p>Lately, in human facial age estimation, local skin features analysis has achieved substantial efficiency [<xref ref-type="bibr" rid="ref-49">49</xref>] as it can represent the significant facial information related to soft tissue aging and remove noise such as hair, background, and non-skin areas. As shown in <xref ref-type="fig" rid="fig-12">Fig. 12</xref>, eleven facial regions were cropped to be used in deriving local texture features using Local Binary Patterns (LBP).</p>
<p>LBP&#x0027;s success stems from the robust binary code that is highly sensitive to texture and soft tissue changes uninfluenced by the light, noise, and facial expressions [<xref ref-type="bibr" rid="ref-18">18</xref>]. LBP takes the value of each pixel as a threshold and calculates 8-bit binary code for their neighboring pixels. Then, a histogram is generated as a texture descriptor [<xref ref-type="bibr" rid="ref-50">50</xref>,<xref ref-type="bibr" rid="ref-51">51</xref>]. The LBP code is expressed as:</p>
<p><disp-formula id="eqn-7">
<label>(7)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-7.png"/><tex-math id="tex-eqn-7"><![CDATA[LB{P_{P,R}} &#x003D; \sum\limits_{P &#x003D; 0}^{P - 1} {s\left( {{g_P} - {g_C}} \right){2^P}} ,\,{\rm where}\,s\left( x \right) &#x003D; \left\{ \matrix{1,x \ge 0 \hfill \cr 0,x < 0 \hfill} \right.]]></tex-math>--><mml:math id="mml-eqn-7" display="block"><mml:mi>L</mml:mi><mml:mi>B</mml:mi><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mo>,</mml:mo><mml:mi>R</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x003D;</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>P</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:munderover><mml:mrow><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>P</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>C</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mn>2</mml:mn><mml:mi>P</mml:mi></mml:msup></mml:mrow></mml:mrow><mml:mo>,</mml:mo><mml:mspace width="thinmathspace"></mml:mspace><mml:mrow><mml:mi mathvariant="normal">w</mml:mi><mml:mi mathvariant="normal">h</mml:mi><mml:mi mathvariant="normal">e</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:mi mathvariant="normal">e</mml:mi></mml:mrow><mml:mspace width="thinmathspace"></mml:mspace><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003D;</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mtable columnspacing="1em" rowspacing="4pt"><mml:mtr><mml:mtd columnalign="left"><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>x</mml:mi><mml:mo>&#x2265;</mml:mo><mml:mn>0</mml:mn></mml:mtd></mml:mtr><mml:mtr><mml:mtd columnalign="left"><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mi>x</mml:mi><mml:mo>&#x003C;</mml:mo><mml:mn>0</mml:mn></mml:mtd></mml:mtr></mml:mtable><mml:mo stretchy="true" symmetric="true" fence="true"></mml:mo></mml:mrow></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-27">
<!--<alternatives><inline-graphic xlink:href="ieqn-27.tif"/><tex-math id="tex-ieqn-27"><![CDATA[{g_C}]]></tex-math>--><mml:math id="mml-ieqn-27"><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>C</mml:mi></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> is the gray value of center pixel,<inline-formula id="ieqn-28">
<!--<alternatives><inline-graphic xlink:href="ieqn-28.tif"/><tex-math id="tex-ieqn-28"><![CDATA[{g_P}]]></tex-math>--><mml:math id="mml-ieqn-28"><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>P</mml:mi></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> is the gray value of equally spaced pixels on the radius circle, <inline-formula id="ieqn-29">
<!--<alternatives><inline-graphic xlink:href="ieqn-29.tif"/><tex-math id="tex-ieqn-29"><![CDATA[s]]></tex-math>--><mml:math id="mml-ieqn-29"><mml:mi>s</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is a thresholding function, <inline-formula id="ieqn-30">
<!--<alternatives><inline-graphic xlink:href="ieqn-30.tif"/><tex-math id="tex-ieqn-30"><![CDATA[P]]></tex-math>--><mml:math id="mml-ieqn-30"><mml:mi>P</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> refers to the number of neighboring pixels and <inline-formula id="ieqn-31">
<!--<alternatives><inline-graphic xlink:href="ieqn-31.tif"/><tex-math id="tex-ieqn-31"><![CDATA[R]]></tex-math>--><mml:math id="mml-ieqn-31"><mml:mi>R</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is the distance between center and neighboring pixels.</p>
<fig id="fig-12">
<label>Figure 12</label>
<caption>
<title>11 skin regions used to extract LBP and Gabor features</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-12.png"/>
</fig>
</sec>
<sec id="s2_3_4">
<label>2.3.4</label>
<title>Gabor Filters</title>
<p>Gabor filters have been commonly used in facial age estimation to extract wrinkles and edge features, because of their ability to determine both the orientation and magnitude of wrinkles and lines [<xref ref-type="bibr" rid="ref-51">51</xref>]. It is characterized by its robustness to noise caused by aspects such as glasses and beard [<xref ref-type="bibr" rid="ref-52">52</xref>]. Here, on each pixel of the face image, filter banks of 2D Gabor filters, including five frequencies and eight orientations, were used in our approach as illustrated in <xref ref-type="fig" rid="fig-13">Fig. 13</xref>. The Gabor filter formula [<xref ref-type="bibr" rid="ref-53">53</xref>] is defined as follows:</p>
<p><disp-formula id="eqn-8">
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-8.png"/><tex-math id="tex-eqn-8"><![CDATA[G\left( {x,y} \right) = \displaystyle{{{f^2}} \over {\pi \gamma \eta }}\exp \left( { - \displaystyle{{{{{x}^{\prime}}^2} + {\gamma ^2}{{{y}^{\prime}}^2}} \over {2{\sigma ^2}}}} \right)\exp \left( {j2\pi f{x}^{\prime} + \phi } \right)]]></tex-math>--><mml:math id="mml-eqn-8" display="block"><mml:mi>G</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003D;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:msup><mml:mi>f</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mrow><mml:mrow><mml:mi>&#x03C0;</mml:mi><mml:mi>&#x03B3;</mml:mi><mml:mi>&#x03B7;</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mo>&#x002B;</mml:mo><mml:mrow><mml:msup><mml:mi>&#x03B3;</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:msup><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mrow><mml:msup><mml:mi>&#x03C3;</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mn>2</mml:mn><mml:mi>&#x03C0;</mml:mi><mml:mi>f</mml:mi><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x002B;</mml:mo><mml:mi>&#x03C6;</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p><disp-formula id="eqn-9">
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-9.png"/><tex-math id="tex-eqn-9"><![CDATA[{x}^{\prime} = x\cos \theta + y\sin \theta]]></tex-math>--><mml:math id="mml-eqn-9" display="block"><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x003D;</mml:mo><mml:mi>x</mml:mi><mml:mi>cos</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mi>&#x03B8;</mml:mi><mml:mo>&#x002B;</mml:mo><mml:mi>y</mml:mi><mml:mi>sin</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mi>&#x03B8;</mml:mi></mml:math>
<!--</alternatives>--></disp-formula></p>
<p><disp-formula id="eqn-10">
<label>(8)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-10.png"/><tex-math id="tex-eqn-10"><![CDATA[{y}^{\prime} = - x\sin \theta - y\cos \theta]]></tex-math>--><mml:math id="mml-eqn-10" display="block"><mml:msup><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x003D;</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mi>x</mml:mi><mml:mi>sin</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mi>&#x03B8;</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>y</mml:mi><mml:mi>cos</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mi>&#x03B8;</mml:mi></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-32">
<!--<alternatives><inline-graphic xlink:href="ieqn-32.tif"/><tex-math id="tex-ieqn-32"><![CDATA[\theta]]></tex-math>--><mml:math id="mml-ieqn-32"><mml:mi>&#x03B8;</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> denotes the orientation, <inline-formula id="ieqn-33">
<!--<alternatives><inline-graphic xlink:href="ieqn-33.tif"/><tex-math id="tex-ieqn-33"><![CDATA[f]]></tex-math>--><mml:math id="mml-ieqn-33"><mml:mi>f</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> represents the frequency, <inline-formula id="ieqn-34">
<!--<alternatives><inline-graphic xlink:href="ieqn-34.tif"/><tex-math id="tex-ieqn-34"><![CDATA[\sigma]]></tex-math>--><mml:math id="mml-ieqn-34"><mml:mi>&#x03C3;</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> refers to standard deviation of Gaussian envelope, <inline-formula id="ieqn-35">
<!--<alternatives><inline-graphic xlink:href="ieqn-35.tif"/><tex-math id="tex-ieqn-35"><![CDATA[\phi]]></tex-math>--><mml:math id="mml-ieqn-35"><mml:mi>&#x03C6;</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is the offset of phase, and <inline-formula id="ieqn-36">
<!--<alternatives><inline-graphic xlink:href="ieqn-36.tif"/><tex-math id="tex-ieqn-36"><![CDATA[\gamma]]></tex-math>--><mml:math id="mml-ieqn-36"><mml:mi>&#x03B3;</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is spatial aspect ratio that specifies the ellipticity to support Gabor function. The local skin regions shown in <xref ref-type="fig" rid="fig-12">Fig. 12</xref> were also used here to extract Gabor features.</p>
<fig id="fig-13">
<label>Figure 13</label>
<caption>
<title>Used 2D Gabor filters in five scales and eight orientations</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-13.png"/>
</fig>
<p>Finally, all computer-vision features were rescaled using Z-score standardization <xref ref-type="disp-formula" rid="eqn-1">Eq. (1)</xref>, then, due to the high dimensionality of shape, appearance, LBP and Gabor features, PCA was performed as a feature subset selection method to determine and select the most effective components. <xref ref-type="table" rid="table-5">Tab. 5</xref> summarizes the determined number of the most effective principal components for each feature type.</p>
<table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Feature dimensions after applying PCA for features selection</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Type of Features</th>
<th>Shape</th>
<th>Appearance</th>
<th>LBP</th>
<th>Gabor</th>
</tr>
</thead>
<tbody>
<tr>
<td>Number of Features</td>
<td>20</td>
<td>200</td>
<td>20</td>
<td>50</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Hierarchical Classification and Regression</title>
<p>Estimating human facial age may be a problem of multi-classification [<xref ref-type="bibr" rid="ref-54">54</xref>&#x2013;<xref ref-type="bibr" rid="ref-56">56</xref>], regression [<xref ref-type="bibr" rid="ref-57">57</xref>,<xref ref-type="bibr" rid="ref-58">58</xref>], or a hybrid of both techniques [<xref ref-type="bibr" rid="ref-18">18</xref>,<xref ref-type="bibr" rid="ref-59">59</xref>]. Many results of previous research explorations have proved that a hybrid or hierarchical age estimation approach outperformed single-stage approaches [<xref ref-type="bibr" rid="ref-60">60</xref>].</p>
<p>In this work, we applied a multi-level hierarchical age estimation approach which classifies images into their appropriate age groups (as high-level age estimation), then predicts an exact age value using regression (as low-level age estimation). A single (one-vs-one) SVM multi-classifier was trained to classify images into one of four age groups: from newborn to toddler (0-3 years), from pre-adolescence to adolescence (4-16 years), from young adult to adult (17-39), and from middle-aged to senior (40-69 Year).</p>
<p>In the regression stage, four SVM regression models were trained. Each regressor was trained using a specific age group to be responsible for predicting ages within the range of this age group. To reduce the effect of misclassification, the regressors were trained with a range overlapping between each adjacent regressors. The range of overlapping among age groups was experimentally selected for better regression performance. <xref ref-type="fig" rid="fig-14">Fig. 14</xref> shows the structure of our multi-level age estimation.</p>
<fig id="fig-14">
<label>Figure 14</label>
<caption>
<title>Multi-level hierarchical classification and regression with age groups overlapping to reduce the effect of misclassification on age prediction</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-14.png"/>
</fig>
</sec>
<sec id="s2_5">
<label>2.5</label>
<title>Evaluation Metrics</title>
<p>The accuracy measure is used to evaluate the performance of SVM in age group classification inferred as:</p>
<p><disp-formula id="eqn-11">
<label>(9)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-11.png"/><tex-math id="tex-eqn-11"><![CDATA[Accuracy = \displaystyle{C \over N} \times 100\%]]></tex-math>--><mml:math id="mml-eqn-11" display="block"><mml:mi>A</mml:mi><mml:mi>c</mml:mi><mml:mi>c</mml:mi><mml:mi>u</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mn>100</mml:mn><mml:mi mathvariant="normal">%</mml:mi></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-37">
<!--<alternatives><inline-graphic xlink:href="ieqn-37.tif"/><tex-math id="tex-ieqn-37"><![CDATA[C]]></tex-math>--><mml:math id="mml-ieqn-37"><mml:mi>C</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is the number of correctly classified images and <inline-formula id="ieqn-38">
<!--<alternatives><inline-graphic xlink:href="ieqn-38.tif"/><tex-math id="tex-ieqn-38"><![CDATA[N]]></tex-math>--><mml:math id="mml-ieqn-38"><mml:mi>N</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> refers to the number of all tested images.</p>
<p>The Mean Absolute Error (MAE) and Cumulative Score (CS) evaluation metrics have been commonly used to evaluate the performance of specific age estimation [<xref ref-type="bibr" rid="ref-61">61</xref>]. MAE is defined as the mean of absolute difference between ground-truth and predicted ages, which is expressed as:</p>
<p><disp-formula id="eqn-12">
<label>(10)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-12.png"/><tex-math id="tex-eqn-12"><![CDATA[MAE = \displaystyle{{\sum\limits_{i = 1}^N {\left| {ag{{\hat e}_i} - ag{e_i}} \right|} } \over N}]]></tex-math>--><mml:math id="mml-eqn-12" display="block"><mml:mi>M</mml:mi><mml:mi>A</mml:mi><mml:mi>E</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mrow><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mrow><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mover><mml:mi>e</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>|</mml:mo></mml:mrow></mml:mrow></mml:mrow><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-39">
<!--<alternatives><inline-graphic xlink:href="ieqn-39.tif"/><tex-math id="tex-ieqn-39"><![CDATA[N]]></tex-math>--><mml:math id="mml-ieqn-39"><mml:mi>N</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> refers to the number of tested images, <inline-formula id="ieqn-40">
<!--<alternatives><inline-graphic xlink:href="ieqn-40.tif"/><tex-math id="tex-ieqn-40"><![CDATA[ag{\hat e_i}]]></tex-math>--><mml:math id="mml-ieqn-40"><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:msub><mml:mrow><mml:mover><mml:mi>e</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> is the predicted age of image <inline-formula id="ieqn-41">
<!--<alternatives><inline-graphic xlink:href="ieqn-41.tif"/><tex-math id="tex-ieqn-41"><![CDATA[i]]></tex-math>--><mml:math id="mml-ieqn-41"><mml:mi>i</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> and <inline-formula id="ieqn-42">
<!--<alternatives><inline-graphic xlink:href="ieqn-42.tif"/><tex-math id="tex-ieqn-42"><![CDATA[ag{e_i}]]></tex-math>--><mml:math id="mml-ieqn-42"><mml:mi>a</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> is the ground-truth age of image <inline-formula id="ieqn-43">
<!--<alternatives><inline-graphic xlink:href="ieqn-43.tif"/><tex-math id="tex-ieqn-43"><![CDATA[i]]></tex-math>--><mml:math id="mml-ieqn-43"><mml:mi>i</mml:mi></mml:math>
<!--</alternatives>--></inline-formula>.</p>
<p>The <italic>CS</italic> is the ratio of the number of images whose errors are less than or equal to the threshold value, divided by the total number of test images. <italic>CS</italic> enables performance comparison at different absolute error levels, which can be defined as:</p>
<p><disp-formula id="eqn-13">
<label>(11)</label>
<!--<alternatives>
<graphic mimetype="image" mime-subtype="png" xlink:href="eqn-13.png"/><tex-math id="tex-eqn-13"><![CDATA[CS = \displaystyle{{{N_{e \le x}}} \over N} \times 100\%]]></tex-math>--><mml:math id="mml-eqn-13" display="block"><mml:mi>C</mml:mi><mml:mi>S</mml:mi><mml:mo>&#x003D;</mml:mo><mml:mstyle scriptlevel="0" displaystyle="true"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>e</mml:mi><mml:mo>&#x2264;</mml:mo><mml:mi>x</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mi>N</mml:mi></mml:mfrac></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mn>100</mml:mn><mml:mi mathvariant="normal">%</mml:mi></mml:mstyle></mml:math>
<!--</alternatives>--></disp-formula></p>
<p>where <inline-formula id="ieqn-44">
<!--<alternatives><inline-graphic xlink:href="ieqn-44.tif"/><tex-math id="tex-ieqn-44"><![CDATA[{N_e}]]></tex-math>--><mml:math id="mml-ieqn-44"><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>e</mml:mi></mml:msub></mml:mrow></mml:math>
<!--</alternatives>--></inline-formula> is the number of images whose estimation error is less than the defined threshold <inline-formula id="ieqn-45">
<!--<alternatives><inline-graphic xlink:href="ieqn-45.tif"/><tex-math id="tex-ieqn-45"><![CDATA[x]]></tex-math>--><mml:math id="mml-ieqn-45"><mml:mi>x</mml:mi></mml:math>
<!--</alternatives>--></inline-formula>, and <inline-formula id="ieqn-46">
<!--<alternatives><inline-graphic xlink:href="ieqn-46.tif"/><tex-math id="tex-ieqn-46"><![CDATA[N]]></tex-math>--><mml:math id="mml-ieqn-46"><mml:mi>N</mml:mi></mml:math>
<!--</alternatives>--></inline-formula> is the total number of test images.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Experiments and Results</title>
<p>Each person in FG-NET dataset has several images at different age stages. Therefore, dividing the dataset into 80% for training and 20% for testing like [<xref ref-type="bibr" rid="ref-62">62</xref>,<xref ref-type="bibr" rid="ref-63">63</xref>], appears to be not an ideal way to test our models, due to the possibility of having very similar images belonging to one person in both training and testing stages. Hence, to prevent that, we have followed a Leave One Person Out (LOPO) validation technique as applied in [<xref ref-type="bibr" rid="ref-64">64</xref>&#x2013;<xref ref-type="bibr" rid="ref-68">68</xref>] to unbiasedly and effectively evaluate our proposed approach. In LOPO, all images belonging to one person are included into testing set while the remaining images of other people in the dataset are used to train the model. As such, the LOPO strategy is repeated as per the number of individuals in the database with a different individual one-by-one used for testing each time, while the final results are calculated from all the estimates.</p>
<table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Accuracy of age group classification using different combinations of vision features, before and after adding semantic features</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Vision-based Features</th>
<th>Before adding<break/>semantic features %</th>
<th>After adding<break/>semantic features %</th>
</tr>
</thead>
<tbody>
<tr>
<td>Ratios &#x002B; Coordinates &#x002B; Appearance</td>
<td>68.36 </td>
<td><bold>69.26 </bold></td>
</tr>
<tr>
<td>Ratios &#x002B; Coordinates &#x002B; LBP</td>
<td>63.87 </td>
<td><bold>75.15</bold></td>
</tr>
<tr>
<td>Ratios &#x002B; Coordinates &#x002B; Gabor</td>
<td>64.37 </td>
<td><bold>71.76 </bold></td>
</tr>
<tr>
<td>Ratios &#x002B; Appearance &#x002B; LBP</td>
<td>67.56 </td>
<td><bold>71.86 </bold></td>
</tr>
<tr>
<td>Ratios &#x002B; Appearance &#x002B; Gabor</td>
<td>65.17 </td>
<td><bold>66.67 </bold></td>
</tr>
<tr>
<td>Ratios &#x002B; LBP &#x002B; Gabor</td>
<td>62.28 </td>
<td><bold>69.16 </bold></td>
</tr>
<tr>
<td>Coordinates &#x002B; Appearance &#x002B; LBP</td>
<td>67.27 </td>
<td><bold>71.36 </bold></td>
</tr>
<tr>
<td>Coordinates &#x002B; Appearance &#x002B; Gabor</td>
<td>65.47 </td>
<td><bold>67.66 </bold></td>
</tr>
<tr>
<td>Coordinates &#x002B; LBP &#x002B; Gabor</td>
<td>64.37 </td>
<td><bold>71.06 </bold></td>
</tr>
<tr>
<td>Appearance &#x002B; LBP &#x002B; Gabor</td>
<td>64.17 </td>
<td><bold>66.17 </bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The accuracy metric is used here to evaluate the performance of SVM in classifying images into their respective age groups. As shown in <xref ref-type="table" rid="table-6">Tab. 6</xref>, we carried out several classifications with various combination of features. The accuracy results were firstly presented using computer-vision-based features alone and secondly after being augmented with our novel semantic features. All proposed 32 semantic features were included in the classification stage based on deduced significance analysis by ANOVA. A significant improvement was observed in all classification experiments after supplanting the vision-based classifiers with the semantic features. The best obtained accuracy result was 75.15% when adding semantic features against only 63.87% when using vision features alone. We will rely on these experiments later to perform regression process.</p>
<p><xref ref-type="fig" rid="fig-15">Fig. 15a</xref> shows the confusion matrix illustrates how images are classified when computer-vision features are used in isolation. It is clear that age group classification is achieved with high misclassification rate, especially in the fourth age group for people of middle-aged to senior (40-69 Year) ages. On the other hand, <xref ref-type="fig" rid="fig-15">Fig. 15b</xref> provides the confusion matrix for the same age group classification approach but after being augmented with the proposed semantic features. As can be observed, the accuracy of age classification was significantly increased in all age groups. Furthermore, it was reported that a majority of misclassified images have been assigned to their adjacent age groups, indicating a closer age group estimation to the actual age group, meaning a smaller range of error unlike the traditional vision-based methods, which were often found randomly assigning a far falsely age group in such misclassification cases. This demonstrates the potency of these semantic features and how they can be capable to distinguish several age groups with less possible error.</p>
<fig id="fig-15">
<label>Figure 15</label>
<caption>
<title>(a) Confusion matrix of age groups classification using vision-based features alone, (b) Confusion matrix of age groups classification using vision-based and semantic features</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-15.png"/>
</fig>
<p>The MAE measurement evaluates the performance of the used regressors in estimating exact ages. In the regression stage, we have eliminated gender, mustache, and beard semantic features from our experiments due to their weak effect and relationship to the exact age value as proved by Pearson&#x0027;s correlation coefficient analysis.</p>
<p><xref ref-type="table" rid="table-7">Tab. 7</xref> displays the results of MAE using various combinations of features. The experiments show that all results improved after combining semantic features with traditional computer-based features. The best MAE we have achieved is 4.41 years old, compared to 5.71 years when semantic features are excluded.</p>
<p><xref ref-type="table" rid="table-8">Tab. 8</xref> summarizes the MAE produced by each age group separately to clarify the positive impact of semantic features on each regressor. The results show how semantic features decrease MAE at all regressors, while a clear improvement was remarkably noticed with older ages within the conducted experiments. This reflects the viability and great sensitivity of our proposed semantic features in capturing and analyzing details of soft facial tissues, which are essential signs of aging in adults and seniors.</p>
<p>The distribution of images in FG-NET dataset is not balanced for all ages, where the number of images decreases as the age increases, since regression models usually need sufficient samples for training to be able to predict ages more efficiently. As a result, the MAE gets worse as the average of ages within the age group increases.</p>
<table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>Mean absolute error (MAE) for exact age prediction before and after adding semantic features</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Vision-based Features</th>
<th>Before adding<break/>semantic features</th>
<th>After adding<break/>semantic features</th>
</tr>
</thead>
<tbody>
<tr>
<td>Ratios &#x002B; Coordinates</td>
<td>6.54</td>
<td><bold>4.66</bold></td>
</tr>
<tr>
<td>Ratios &#x002B; Appearance</td>
<td>5.71</td>
<td><bold>4.41</bold></td>
</tr>
<tr>
<td>Ratios &#x002B; LBP</td>
<td>6.41</td>
<td><bold>4.58</bold></td>
</tr>
<tr>
<td>Ratios &#x002B; Gabor</td>
<td>6.65</td>
<td><bold>5.02</bold></td>
</tr>
<tr>
<td>Coordinates &#x002B; Appearance</td>
<td>5.64</td>
<td><bold>4.53</bold></td>
</tr>
<tr>
<td>Coordinates &#x002B; LBP</td>
<td>6.45</td>
<td><bold>4.62</bold></td>
</tr>
<tr>
<td>Coordinates &#x002B; Gabor</td>
<td>6.53</td>
<td><bold>5.19</bold></td>
</tr>
<tr>
<td>Appearance &#x002B; LBP</td>
<td>5.84</td>
<td><bold>4.71</bold></td>
</tr>
<tr>
<td>Appearance &#x002B; Gabor</td>
<td>6.44</td>
<td><bold>5.52</bold></td>
</tr>
<tr>
<td>LBP &#x002B; Gabor</td>
<td>7.12</td>
<td><bold>5.61</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="table-8">
<label>Table 8</label>
<caption>
<title>(MAE) for each regressor separately and their overall MAE</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Age estimation performance</th>
<th>MAE 1</th>
<th>MAE 2</th>
<th>MAE 3</th>
<th>MAE 4</th>
<th>Overall MAE</th>
</tr>
</thead>
<tbody>
<tr>
<td>Without Semantic Features</td>
<td>3.18</td>
<td>4.59</td>
<td>7.38</td>
<td>14.93</td>
<td>5.71</td>
</tr>
<tr>
<td>With Semantic Features</td>
<td><bold>2.03</bold></td>
<td><bold>3.55</bold></td>
<td><bold>5.91</bold></td>
<td><bold>8.95</bold></td>
<td><bold>4.41</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The cumulative score was computed at different error levels ranging from 0 to10 years. Consequently, 83.53% of images have a level error less than or equal to 10 years when using computer-vision features, while the cumulative score percentage improved to reach 90.12% after augmenting them with semantic features.</p>
<p>After a considerable improvement was proved for the new proposed semantic features by all means in the carried out experiments, we have used them alone to train a classifier and regressors on age estimation task for validating their pure viability and performance in isolation. The received results were 61.88% classification accuracy, 6.15 MAE, and 83.23% CS for age prediction. <xref ref-type="fig" rid="fig-16">Fig. 16</xref> shows and compares the CS performance when using computer-vision features, semantic features, and the fusion of both semantic and vision features.</p>
<fig id="fig-16">
<label>Figure 16</label>
<caption>
<title>Cumulative Score of computer-vision features, semantic features, and augmenting computer-vision with semantic features</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-16.png"/>
</fig>
<p>Finally, we used the features of the second experiment as shown in <xref ref-type="table" rid="table-7">Tab. 7</xref> that achieves the best MAE (4.41 years); a Bayesian optimization was performed for each regressor independently, because each of them deals with a different age range. The Bayesian optimization intends to search for optimum hyperparameters that give a minimum error. We obtained 3.90 MAE and 92.42% CS at 10 years&#x2019; error level. <xref ref-type="fig" rid="fig-17">Fig. 17</xref> and <xref ref-type="table" rid="table-9">Tab. 9</xref> compare the CS and MAE before and after optimization process, respectively. <xref ref-type="fig" rid="fig-18">Fig. 18</xref> provides samples of good and poor age estimation of some images along with their respective actual and estimated age value.</p>
<table-wrap id="table-9">
<label>Table 9</label>
<caption>
<title>The best MAE obtained before and after tuning the hyperparameters of each regressors</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Proposed Approach Performance</th>
<th>MAE 1</th>
<th>MAE 2</th>
<th>MAE 3</th>
<th>MAE 4</th>
<th>Overall MAE</th>
</tr>
</thead>
<tbody>
<tr>
<td>Before Optimization</td>
<td>2.03</td>
<td>3.55</td>
<td>5.91</td>
<td>8.95</td>
<td>4.41</td>
</tr>
<tr>
<td>After Optimization</td>
<td><bold>1.31</bold></td>
<td><bold>3.06</bold></td>
<td><bold>5.62</bold></td>
<td><bold>7.75</bold></td>
<td><bold>3.90</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="fig-17">
<label>Figure 17</label>
<caption>
<title>Cumulative Score of our proposed augmented approach before and after optimization process</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-17.png"/>
</fig>
<fig id="fig-18">
<label>Figure 18</label>
<caption>
<title>Sample results of different age estimations; (a) provides good estimations, (b) provides poor estimations</title>
</caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_19533-fig-18.png"/>
</fig>
</sec>
<sec id="s4">
<label>4</label>
<title>Conclusion and Future Work</title>
<p>In this paper, we investigated the problem of human age estimation from face images. We proposed a set of novel semantic features representing various facial aging characteristics or signs. We examine significance and effectiveness of each semantic feature using diverse statistical analysis tests, then we exploit powerful features to augment traditional computer-vision-based features. A hierarchical estimation approach was applied by performing classification to predict age group followed by regression to estimate the exact age. FG-NET aging dataset with variety in expressions, races, resolutions, poses and illumination was used to evaluate the performance of our proposed approach.</p>
<p>Experimental results proved a significant and remarkable improvement in age estimation performance, by all means, occurred at the results of both stages: age group (high-level) classification; and exact age (low-level) prediction, when fusing and supplementing traditional computer vision-based features with our proposed human-based semantic features, especially with ages in the fourth age group ranging from middle-aged to senior (<italic>i.e</italic>., from 40 to 69 years old).</p>
<p>This initial research study may establish further research tracks to be undertaken by researchers in the future utilizing our obtained results and findings in how to supplement computer vision-based features with human-based features for augmented facial-based human age estimation.</p>
<p><bold>As future work</bold>, increased discrimination for more robust age estimation capabilities can be investigated by fusing our proposed semantic features with deep features, where it is highly expected, based on this research&#x2019;s results and conclusions, that our semantic features would be most likely capable to augment/enhance the performance when fused with the most recent models based on deep learning and CNN, and to outperform their performance when used in isolation. Learning for automatic semantic face feature extraction and annotation is a potential desired future work that helps researchers to improve computer-vision techniques and bridge the semantic gap between machines and humans automatically.</p>
</sec>
</body>
<back>
<ack>
<p>The authors would like to thank King Abdulaziz University Scientific Endowment for funding the research reported in this paper.</p>
</ack><fn-group>
<fn fn-type="other">
<p><bold>Funding Statement:</bold> This research was supported and funded by KAU Scientific Endowment, King Abdulaziz University, Jeddah, Saudi Arabia.</p>
</fn>
<fn fn-type="conflict">
<p><bold>Conflicts of Interest:</bold> The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</fn>
</fn-group>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1">
<label>1</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>M. C.</given-names> 
<surname>Voelkle</surname></string-name>, <string-name>
<given-names>N. C.</given-names> 
<surname>Ebner</surname></string-name>, <string-name>
<given-names>U.</given-names> 
<surname>Lindenberger</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Riediger</surname></string-name>
</person-group>, &#x201C;
<article-title>Let me guess how old you are: Effects of age, gender, and facial expression on perceptions of age</article-title>,&#x201D; 
<source>Psychology and Aging</source>, vol. 
<volume>27</volume>, no. 
<issue>2</issue>, pp. 
<fpage>265</fpage>&#x2013;
<lpage>277</lpage>, 
<year>2012</year>.</mixed-citation>
</ref>
<ref id="ref-2">
<label>2</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>A. A.</given-names> 
<surname>Shejul</surname></string-name>, <string-name>
<given-names>K. S.</given-names> 
<surname>Kinage</surname></string-name> and <string-name>
<given-names>B. E.</given-names> 
<surname>Reddy</surname></string-name>
</person-group>, &#x201C;
<article-title>Comprehensive review on facial based human age estimation</article-title>,&#x201D; in <conf-name>2017 Int. Conf. on Energy, Communication, Data Analytics and Soft Computing</conf-name>, 
<publisher-loc>Chennai, India</publisher-loc>, pp. 
<fpage>3211</fpage>&#x2013;
<lpage>3216</lpage>, 
<year>2017</year>. </mixed-citation>
</ref>
<ref id="ref-3">
<label>3</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>M.</given-names> 
<surname>Georgopoulos</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Panagakis</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Pantic</surname></string-name>
</person-group>, &#x201C;
<article-title>Modeling of facial aging and kinship: A survey</article-title>,&#x201D; 
<source>Image and Vision Computing</source>, vol. 
<volume>80</volume>, no. 
<issue>14</issue>, pp. 
<fpage>58</fpage>&#x2013;
<lpage>79</lpage>, 
<year>2018</year>.</mixed-citation>
</ref>
<ref id="ref-4">
<label>4</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>R.</given-names> 
<surname>Angulu</surname></string-name>, <string-name>
<given-names>J. R.</given-names> 
<surname>Tapamo</surname></string-name> and <string-name>
<given-names>A. O.</given-names> 
<surname>Adewumi</surname></string-name>
</person-group>, &#x201C;
<article-title>Age estimation via face images: A survey</article-title>,&#x201D; 
<source>EURASIP Journal on Image and Video Processing</source>, vol. 
<volume>2018</volume>, no. 
<issue>1</issue>, pp. 
<fpage>179</fpage>, 
<year>2018</year>.</mixed-citation>
</ref>
<ref id="ref-5">
<label>5</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>M.</given-names> 
<surname>Kaur</surname></string-name>, <string-name>
<given-names>R. K.</given-names> 
<surname>Garg</surname></string-name> and <string-name>
<given-names>S.</given-names> 
<surname>Singla</surname></string-name>
</person-group>, &#x201C;
<article-title>Analysis of facial soft tissue changes with aging and their effects on facial morphology: A forensic perspective</article-title>,&#x201D; 
<source>Egyptian Journal of Forensic Sciences</source>, vol. 
<volume>5</volume>, no. 
<issue>2</issue>, pp. 
<fpage>46</fpage>&#x2013;
<lpage>56</lpage>, 
<year>2015</year>.</mixed-citation>
</ref>
<ref id="ref-6">
<label>6</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A.</given-names> 
<surname>Othmani</surname></string-name>, <string-name>
<given-names>A. R.</given-names> 
<surname>Taleb</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Abdelkawy</surname></string-name> and <string-name>
<given-names>A.</given-names> 
<surname>Hadid</surname></string-name>
</person-group>, &#x201C;
<article-title>Age estimation from faces using deep learning: A comparative analysis</article-title>,&#x201D; 
<source>Computer Vision and Image Understanding</source>, vol. 
<volume>196</volume>, no. 
<issue>8</issue>, pp. 
<fpage>102961</fpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-7">
<label>7</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>N.</given-names> 
<surname>Ramanathan</surname></string-name> and <string-name>
<given-names>R.</given-names> 
<surname>Chellappa</surname></string-name>
</person-group>, &#x201C;
<article-title>Modeling age progression in young faces</article-title>,&#x201D; in <conf-name>2006 IEEE Computer Society Conf. on Computer Vision and Pattern Recognition (CVPR&#x0027;06)</conf-name>, 
<publisher-loc>New York, NY, USA</publisher-loc>, pp. 
<fpage>387</fpage>&#x2013;
<lpage>394</lpage>, 
<year>2006</year>. </mixed-citation>
</ref>
<ref id="ref-8">
<label>8</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>T.</given-names> 
<surname>Dhimar</surname></string-name> and <string-name>
<given-names>K.</given-names> 
<surname>Mistree</surname></string-name>
</person-group>, &#x201C;
<article-title>Feature extraction for facial age estimation: A survey</article-title>,&#x201D; in <conf-name>2016 Int. Conf. on Wireless Communications, Signal Processing and Networking (WiSPNET)</conf-name>, 
<publisher-loc>Chennai, India</publisher-loc>, pp. 
<fpage>2243</fpage>&#x2013;
<lpage>2248</lpage>, 
<year>2016</year>. </mixed-citation>
</ref>
<ref id="ref-9">
<label>9</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S.</given-names> 
<surname>Kumar</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Ranjitha</surname></string-name> and <string-name>
<given-names>H.</given-names> 
<surname>Suresh</surname></string-name>
</person-group>, &#x201C;
<article-title>An active age estimation of facial image using anthropometric model and fast ica</article-title>,&#x201D; 
<source>Journal of Engineering Science &#x0026; Technology Review</source>, vol. 
<volume>10</volume>, no. 
<issue>1, </issue>
<year>2017</year>.</mixed-citation>
</ref>
<ref id="ref-10">
<label>10</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>F. B.</given-names> 
<surname>Alvi</surname></string-name>, <string-name>
<given-names>R.</given-names> 
<surname>Pears</surname></string-name> and <string-name>
<given-names>N.</given-names> 
<surname>Kasabov</surname></string-name>
</person-group>, &#x201C;
<article-title>An evolving spatio-temporal approach for gender and age group classification with spiking neural networks</article-title>,&#x201D; 
<source>Evolving Systems</source>, vol. 
<volume>9</volume>, no. 
<issue>2</issue>, pp. 
<fpage>145</fpage>&#x2013;
<lpage>156</lpage>, 
<year>2018</year>.</mixed-citation>
</ref>
<ref id="ref-11">
<label>11</label><mixed-citation publication-type="other">
<person-group person-group-type="author"><string-name>
<given-names>L. F.</given-names> 
<surname>Porto</surname></string-name>, <string-name>
<given-names>L. N. C.</given-names> 
<surname>Lima</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Franco</surname></string-name>, <string-name>
<given-names>D. M.</given-names> 
<surname>Pianto</surname></string-name>, <string-name>
<given-names>C. E. M.</given-names> 
<surname>Palhares</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>Estimating sex and age for forensic applications using machine learning based on facial measurements from frontal cephalometric landmarks</article-title>,&#x201D; 
<source>arXiv preprint arXiv:1908.02353,</source> 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-12">
<label>12</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>P.</given-names> 
<surname>Thukral</surname></string-name>, <string-name>
<given-names>K.</given-names> 
<surname>Mitra</surname></string-name> and <string-name>
<given-names>R.</given-names> 
<surname>Chellappa</surname></string-name>
</person-group>, &#x201C;
<article-title>A hierarchical approach for human age estimation</article-title>,&#x201D; in <conf-name>2012 IEEE Int. Conf. on Acoustics, Speech and Signal Processing</conf-name>, 
<publisher-loc>Kyoto, Japan</publisher-loc>, pp. 
<fpage>1529</fpage>&#x2013;
<lpage>1532</lpage>, 
<year>2012</year>. </mixed-citation>
</ref>
<ref id="ref-13">
<label>13</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>M.</given-names> 
<surname>Iqtait</surname></string-name>, <string-name>
<given-names>F.</given-names> 
<surname>Mohamad</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Mamat</surname></string-name>
</person-group>, &#x201C;
<article-title>Feature extraction for face recognition via active shape model (asm) and active appearance model (aam)</article-title>,&#x201D; in <conf-name>IOP Conf. Series: Materials Science and Engineering</conf-name>, 
<publisher-loc>Tangerang Selatan, Indonesia</publisher-loc>, pp. 
<fpage>12032</fpage>, 
<year>2018</year>. </mixed-citation>
</ref>
<ref id="ref-14">
<label>14</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S. A.</given-names> 
<surname>Rizwan</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Jalal</surname></string-name>, <string-name>
<given-names>M.</given-names> 
<surname>Gochoo</surname></string-name> and <string-name>
<given-names>K.</given-names> 
<surname>Kim</surname></string-name>
</person-group>, &#x201C;
<article-title>Robust active shape model via hierarchical feature extraction with sfs-optimized convolution neural network for invariant human age classification</article-title>,&#x201D; 
<source>Electronics</source>, vol. 
<volume>10</volume>, no. 
<issue>4</issue>, pp. 
<fpage>465</fpage>, 
<year>2021</year>.</mixed-citation>
</ref>
<ref id="ref-15">
<label>15</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>P. C. S.</given-names> 
<surname>Reddy</surname></string-name>, <string-name>
<given-names>K.</given-names> 
<surname>Sarma</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Sharma</surname></string-name>, <string-name>
<given-names>P. V.</given-names> 
<surname>Rao</surname></string-name>, <string-name>
<given-names>S. G.</given-names> 
<surname>Rao</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;<article-title>Enhanced age prediction and gender classification (eap-gc) framework using regression and svm techniques</article-title>,&#x201D; 
<article-title>Materials Today: Proc.</article-title>, 
<year>2020</year>. </mixed-citation>
</ref>
<ref id="ref-16">
<label>16</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A. A.</given-names> 
<surname>Shejul</surname></string-name>, <string-name>
<given-names>K. S.</given-names> 
<surname>Kinage</surname></string-name> and <string-name>
<given-names>B. E.</given-names> 
<surname>Reddy</surname></string-name>
</person-group>, &#x201C;
<article-title>Cdbn: Crow deep belief network based on scattering and aam features for age estimation</article-title>,&#x201D; 
<source>Journal of Signal Processing Systems</source>, vol. 
<volume>92</volume>, no. 
<issue>10</issue>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>19</lpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-17">
<label>17</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>V.</given-names> 
<surname>Martin</surname></string-name>, <string-name>
<given-names>R.</given-names> 
<surname>Seguier</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Porcheron</surname></string-name> and <string-name>
<given-names>F.</given-names> 
<surname>Morizot</surname></string-name>
</person-group>, &#x201C;
<article-title>Face aging simulation with a new wrinkle oriented active appearance model</article-title>,&#x201D; 
<source>Multimedia Tools and Applications</source>, vol. 
<volume>78</volume>, no. 
<issue>5</issue>, pp. 
<fpage>6309</fpage>&#x2013;
<lpage>6327</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-18">
<label>18</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>T.</given-names> 
<surname>Wu</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Zhao</surname></string-name>, <string-name>
<given-names>L.</given-names> 
<surname>Liu</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Li</surname></string-name>, <string-name>
<given-names>W.</given-names> 
<surname>Xu</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>A novel hierarchical regression approach for human facial age estimation based on deep forest</article-title>,&#x201D; in <conf-name>2018 IEEE 15th Int. Conf. on Networking, Sensing and Control</conf-name>, 
<publisher-loc>Zhuhai, China</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>6</lpage>, 
<year>2018</year>. </mixed-citation>
</ref>
<ref id="ref-19">
<label>19</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>J. K.</given-names> 
<surname>Pontes</surname></string-name>, <string-name>
<given-names>A. S.</given-names> 
<surname>Britto</surname> 
<suffix>Jr</suffix></string-name>, <string-name>
<given-names>C.</given-names> 
<surname>Fookes</surname></string-name> and <string-name>
<given-names>A. L.</given-names> 
<surname>Koerich</surname></string-name>
</person-group>, &#x201C;
<article-title>A flexible hierarchical approach for facial age estimation based on multiple features</article-title>,&#x201D; 
<source>Pattern Recognition</source>, vol. 
<volume>54</volume>, no. 
<issue>10</issue>, pp. 
<fpage>34</fpage>&#x2013;
<lpage>51</lpage>, 
<year>2016</year>.</mixed-citation>
</ref>
<ref id="ref-20">
<label>20</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>X.</given-names> 
<surname>Geng</surname></string-name>, <string-name>
<given-names>Z.-H.</given-names> 
<surname>Zhou</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Zhang</surname></string-name>, <string-name>
<given-names>G.</given-names> 
<surname>Li</surname></string-name> and <string-name>
<given-names>H.</given-names> 
<surname>Dai</surname></string-name>
</person-group>, &#x201C;
<article-title>Learning from facial aging patterns for automatic age estimation</article-title>,&#x201D; in <conf-name>Proc. of the 14th ACM Int. Conf. on Multimedia</conf-name>, 
<publisher-loc>Santa Barbara, CA, USA</publisher-loc>, pp. 
<fpage>307</fpage>&#x2013;
<lpage>316</lpage>, 
<year>2006</year>. </mixed-citation>
</ref>
<ref id="ref-21">
<label>21</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>T. K.</given-names> 
<surname>Sahoo</surname></string-name> and <string-name>
<given-names>H.</given-names> 
<surname>Banka</surname></string-name>
</person-group>, &#x201C;
<article-title>Multi-feature-based facial age estimation using an incomplete facial aging database</article-title>,&#x201D; 
<source>Arabian Journal for Science and Engineering</source>, vol. 
<volume>43</volume>, no. 
<issue>12</issue>, pp. 
<fpage>8057</fpage>&#x2013;
<lpage>8078</lpage>, 
<year>2018</year>.</mixed-citation>
</ref>
<ref id="ref-22">
<label>22</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>P.</given-names> 
<surname>Punyani</surname></string-name>, <string-name>
<given-names>R.</given-names> 
<surname>Gupta</surname></string-name> and <string-name>
<given-names>A.</given-names> 
<surname>Kumar</surname></string-name>
</person-group>, &#x201C;
<article-title>Neural networks for facial age estimation: A survey on recent advances</article-title>,&#x201D; 
<source>Artificial Intelligence Review</source>, vol. 
<volume>53</volume>, no. 
<issue>5</issue>, pp. 
<fpage>3299</fpage>&#x2013;
<lpage>3347</lpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-23">
<label>23</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S.</given-names> 
<surname>Taheri</surname></string-name> and <string-name>
<given-names>&#x00D6;.</given-names> 
<surname>Toygar</surname></string-name>
</person-group>, &#x201C;
<article-title>On the use of dag-cnn architecture for age estimation with multi-stage features fusion</article-title>,&#x201D; 
<source>Neurocomputing</source>, vol. 
<volume>329</volume>, no. 
<issue>1</issue>, pp. 
<fpage>300</fpage>&#x2013;
<lpage>310</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-24">
<label>24</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S.</given-names> 
<surname>Taheri</surname></string-name> and <string-name>
<given-names>&#x00D6;.</given-names> 
<surname>Toygar</surname></string-name>
</person-group>, &#x201C;
<article-title>Multi-stage age estimation using two level fusions of handcrafted and learned features on facial images</article-title>,&#x201D; 
<source>IET Biometrics</source>, vol. 
<volume>8</volume>, no. 
<issue>2</issue>, pp. 
<fpage>124</fpage>&#x2013;
<lpage>133</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-25">
<label>25</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>C.</given-names> 
<surname>Szegedy</surname></string-name>, <string-name>
<given-names>W.</given-names> 
<surname>Liu</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Jia</surname></string-name>, <string-name>
<given-names>P.</given-names> 
<surname>Sermanet</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Reed</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>Going deeper with convolutions</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, 
<publisher-loc>Boston, Massachusetts</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>9</lpage>, 
<year>2015</year>. </mixed-citation>
</ref>
<ref id="ref-26">
<label>26</label><mixed-citation publication-type="other">
<person-group person-group-type="author"><string-name>
<given-names>K.</given-names> 
<surname>Simonyan</surname></string-name> and <string-name>
<given-names>A.</given-names> 
<surname>Zisserman</surname></string-name>
</person-group>, &#x201C;
<article-title>Very deep convolutional networks for large-scale image recognition</article-title>,&#x201D; 
<comment>arXiv preprint arXiv: 1409. 1556</comment>, 
<year>2014</year>.</mixed-citation>
</ref>
<ref id="ref-27">
<label>27</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>P.</given-names> 
<surname>Smith</surname></string-name> and <string-name>
<given-names>C.</given-names> 
<surname>Chen</surname></string-name>
</person-group>, &#x201C;
<article-title>Transfer learning with deep cnns for gender recognition and age estimation</article-title>,&#x201D; in <conf-name>2018 IEEE Int. Conf. on Big Data (Big Data)</conf-name>, 
<publisher-loc>Seattle, WA, USA</publisher-loc>, pp. 
<fpage>2564</fpage>&#x2013;
<lpage>2571</lpage>, 
<year>2018</year>. </mixed-citation>
</ref>
<ref id="ref-28">
<label>28</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>H.</given-names> 
<surname>Han</surname></string-name>, <string-name>
<given-names>C.</given-names> 
<surname>Otto</surname></string-name> and <string-name>
<given-names>A. K.</given-names> 
<surname>Jain</surname></string-name>
</person-group>, &#x201C;
<article-title>Age estimation from face images: Human vs. Machine performance</article-title>,&#x201D; in <conf-name>2013 Int. Conf. on Biometrics</conf-name>, 
<publisher-loc>Madrid, Spain</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>8</lpage>, 
<year>2013</year>. </mixed-citation>
</ref>
<ref id="ref-29">
<label>29</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>H.</given-names> 
<surname>Han</surname></string-name>, <string-name>
<given-names>C.</given-names> 
<surname>Otto</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Liu</surname></string-name> and <string-name>
<given-names>A. K.</given-names> 
<surname>Jain</surname></string-name>
</person-group>, &#x201C;
<article-title>Demographic estimation from face images: Human vs. Machine performance</article-title>,&#x201D; 
<source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>, vol. 
<volume>37</volume>, no. 
<issue>6</issue>, pp. 
<fpage>1148</fpage>&#x2013;
<lpage>1161</lpage>, 
<year>2015</year>.</mixed-citation>
</ref>
<ref id="ref-30">
<label>30</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>J.</given-names> 
<surname>Alarifi</surname></string-name>, <string-name>
<given-names>J.</given-names> 
<surname>Fry</surname></string-name>, <string-name>
<given-names>D.</given-names> 
<surname>Dancey</surname></string-name> and <string-name>
<given-names>M. H.</given-names> 
<surname>Yap</surname></string-name>
</person-group>, &#x201C;
<article-title>Understanding face age estimation: Humans and machine</article-title>,&#x201D; in <conf-name>2019 Int. Conf. on Computer, Information and Telecommunication Systems</conf-name>, 
<publisher-loc>Beijing, China</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>5</lpage>, 
<year>2019</year>. </mixed-citation>
</ref>
<ref id="ref-31">
<label>31</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>E. S.</given-names> 
<surname>Jaha</surname></string-name>
</person-group>, &#x201C;
<article-title>Augmenting gabor-based face recognition with global soft biometrics</article-title>,&#x201D; in <conf-name>2019 7th Int. Sym. on Digital Forensics and Security</conf-name>, 
<publisher-loc>Barcelos, Portugal</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>5</lpage>, 
<year>2019</year>. </mixed-citation>
</ref>
<ref id="ref-32">
<label>32</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>E. S.</given-names> 
<surname>Jaha</surname></string-name> and <string-name>
<given-names>M. S.</given-names> 
<surname>Nixon</surname></string-name>
</person-group>, &#x201C;
<article-title>Soft biometrics for subject identification using clothing attributes</article-title>,&#x201D; in <conf-name>IEEE Int. Joint Conf. on Biometrics</conf-name>, 
<publisher-loc>Clearwater, FL, USA</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>6</lpage>, 
<year>2014</year>. </mixed-citation>
</ref>
<ref id="ref-33">
<label>33</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S.</given-names> 
<surname>Samangooei</surname></string-name> and <string-name>
<given-names>M. S.</given-names> 
<surname>Nixon</surname></string-name>
</person-group>, &#x201C;
<article-title>Performing content-based retrieval of humans using gait biometrics</article-title>,&#x201D; 
<source>Multimedia Tools and Applications</source>, vol. 
<volume>49</volume>, no. 
<issue>1</issue>, pp. 
<fpage>195</fpage>&#x2013;
<lpage>212</lpage>, 
<year>2010</year>.</mixed-citation>
</ref>
<ref id="ref-34">
<label>34</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>G.</given-names> 
<surname>Panis</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Lanitis</surname></string-name>, <string-name>
<given-names>N.</given-names> 
<surname>Tsapatsoulis</surname></string-name> and <string-name>
<given-names>T. F.</given-names> 
<surname>Cootes</surname></string-name>
</person-group>, &#x201C;
<article-title>Overview of research on facial ageing using the FG-NET ageing database</article-title>,&#x201D; 
<source>IET Biometrics</source>, vol. 
<volume>5</volume>, no. 
<issue>2</issue>, pp. 
<fpage>37</fpage>&#x2013;
<lpage>46</lpage>, 
<year>2016</year>.</mixed-citation>
</ref>
<ref id="ref-35">
<label>35</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>V.</given-names> 
<surname>Carletti</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Greco</surname></string-name>, <string-name>
<given-names>G.</given-names> 
<surname>Percannella</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Vento</surname></string-name>
</person-group>, &#x201C;
<article-title>Age from faces in the deep learning revolution</article-title>,&#x201D; 
<source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>, vol. 
<volume>42</volume>, no. 
<issue>9</issue>, pp. 
<fpage>2113</fpage>&#x2013;
<lpage>2132</lpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-36">
<label>36</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>O. F.</given-names> 
<surname>Osman</surname></string-name> and <string-name>
<given-names>M. H.</given-names> 
<surname>Yap</surname></string-name>
</person-group>, &#x201C;
<article-title>Computational intelligence in automatic face age estimation: A survey</article-title>,&#x201D; 
<source>IEEE Transactions on Emerging Topics in Computational Intelligence</source>, vol. 
<volume>3</volume>, no. 
<issue>3</issue>, pp. 
<fpage>271</fpage>&#x2013;
<lpage>285</lpage>, 
<year>2019</year>.</mixed-citation>
</ref>
<ref id="ref-37">
<label>37</label><mixed-citation publication-type="other">
<article-title>Figure eight</article-title>, 
<year>2020</year>. 
<comment>Available:</comment> <uri>https://appen.com/figure-eight-is-now-appen/</uri>.</mixed-citation>
</ref>
<ref id="ref-38">
<label>38</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A. M.</given-names> 
<surname>Albert</surname></string-name>, <string-name>
<given-names>K.</given-names> 
<surname>Ricanek</surname> 
<suffix>Jr</suffix></string-name> and <string-name>
<given-names>E.</given-names> 
<surname>Patterson</surname></string-name>
</person-group>, &#x201C;
<article-title>A review of the literature on the aging adult skull and face: Implications for forensic science research and applications</article-title>,&#x201D; 
<source>Forensic Science International</source>, vol. 
<volume>172</volume>, no. 
<issue>1</issue>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>9</lpage>, 
<year>2007</year>.</mixed-citation>
</ref>
<ref id="ref-39">
<label>39</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S. R.</given-names> 
<surname>Coleman</surname></string-name> and <string-name>
<given-names>R.</given-names> 
<surname>Grover</surname></string-name>
</person-group>, &#x201C;
<article-title>The anatomy of the aging face: Volume loss and changes in 3-dimensional topography</article-title>,&#x201D; 
<source>Aesthetic Surgery Journal</source>, vol. 
<volume>26</volume>, no. 
<issue>1_Supplement</issue>, pp. 
<fpage>S4</fpage>&#x2013;
<lpage>S9</lpage>, 
<year>2006</year>.</mixed-citation>
</ref>
<ref id="ref-40">
<label>40</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A.</given-names> 
<surname>Porcheron</surname></string-name>, <string-name>
<given-names>E.</given-names> 
<surname>Mauger</surname></string-name> and <string-name>
<given-names>R.</given-names> 
<surname>Russell</surname></string-name>
</person-group>, &#x201C;
<article-title>Aspects of facial contrast decrease with age and are cues for age perception</article-title>,&#x201D; 
<source>PLoS One</source>, vol. 
<volume>8</volume>, no. 
<issue>3</issue>, pp. 
<fpage>e57985</fpage>, 
<year>2013</year>.</mixed-citation>
</ref>
<ref id="ref-41">
<label>41</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>K. H.</given-names> 
<surname>Liu</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Yan</surname></string-name> and <string-name>
<given-names>C. C. J.</given-names> 
<surname>Kuo</surname></string-name>
</person-group>, &#x201C;
<article-title>Age estimation via grouping and decision fusion</article-title>,&#x201D; 
<source>IEEE Transactions on Information Forensics and Security</source>, vol. 
<volume>10</volume>, no. 
<issue>11</issue>, pp. 
<fpage>2408</fpage>&#x2013;
<lpage>2423</lpage>, 
<year>2015</year>.</mixed-citation>
</ref>
<ref id="ref-42">
<label>42</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>C. E. P.</given-names> 
<surname>Machado</surname></string-name>, <string-name>
<given-names>M. R. P.</given-names> 
<surname>Flores</surname></string-name>, <string-name>
<given-names>L. N. C.</given-names> 
<surname>Lima</surname></string-name>, <string-name>
<given-names>R. L. R.</given-names> 
<surname>Tinoco</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Franco</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>A new approach for the analysis of facial growth and age estimation: Iris ratio</article-title>,&#x201D; 
<source>PLOS One</source>, vol. 
<volume>12</volume>, no. 
<issue>7</issue>, pp. 
<fpage>e0180330</fpage>, 
<year>2017</year>.</mixed-citation>
</ref>
<ref id="ref-43">
<label>43</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>M.</given-names> 
<surname>Alamri</surname></string-name> and <string-name>
<given-names>S.</given-names> 
<surname>Mahmoodi</surname></string-name>
</person-group>, &#x201C;
<article-title>Facial profiles recognition using comparative facial soft biometrics</article-title>,&#x201D; in <conf-name>2020 Int. Conf. of the Biometrics Special Interest Group</conf-name>, 
<publisher-loc>Darmstadt, Germany</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>4</lpage>, 
<year>2020</year>. </mixed-citation>
</ref>
<ref id="ref-44">
<label>44</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>T. F.</given-names> 
<surname>Cootes</surname></string-name>, <string-name>
<given-names>G. J.</given-names> 
<surname>Edwards</surname></string-name> and <string-name>
<given-names>C. J.</given-names> 
<surname>Taylor</surname></string-name>
</person-group>, &#x201C;
<article-title>Active appearance models</article-title>,&#x201D; in <conf-name>European Conf. on Computer Vision</conf-name>, 
<publisher-loc>Berlin, Heidelberg</publisher-loc>, pp. 
<fpage>484</fpage>&#x2013;
<lpage>498</lpage>, 
<year>1998</year>. </mixed-citation>
</ref>
<ref id="ref-45">
<label>45</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>X.</given-names> 
<surname>Gao</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Su</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Li</surname></string-name> and <string-name>
<given-names>D.</given-names> 
<surname>Tao</surname></string-name>
</person-group>, &#x201C;
<article-title>A review of active appearance models</article-title>,&#x201D; 
<source>IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews)</source>, vol. 
<volume>40</volume>, no. 
<issue>2</issue>, pp. 
<fpage>145</fpage>&#x2013;
<lpage>158</lpage>, 
<year>2010</year>.</mixed-citation>
</ref>
<ref id="ref-46">
<label>46</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>G.</given-names> 
<surname>Tzimiropoulos</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Pantic</surname></string-name>
</person-group>, &#x201C;
<article-title>Optimization problems for fast aam fitting in-the-wild</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Int. Conf. on Computer Vision</conf-name>, 
<publisher-loc>Sydney, NSW, Australia</publisher-loc>, pp. 
<fpage>593</fpage>&#x2013;
<lpage>600</lpage>, 
<year>2013</year>. </mixed-citation>
</ref>
<ref id="ref-47">
<label>47</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>Y.</given-names> 
<surname>Liang</surname></string-name>, <string-name>
<given-names>X.</given-names> 
<surname>Wang</surname></string-name>, <string-name>
<given-names>L.</given-names> 
<surname>Zhang</surname></string-name> and <string-name>
<given-names>Z.</given-names> 
<surname>Wang</surname></string-name>
</person-group>, &#x201C;
<article-title>A hierarchical framework for facial age estimation</article-title>,&#x201D; 
<source>Mathematical Problems in Engineering</source>, vol. 
<volume>2014</volume>, no. 
<issue>6</issue>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>8</lpage>, 
<year>2014</year>.</mixed-citation>
</ref>
<ref id="ref-48">
<label>48</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>P.</given-names> 
<surname>Koruga</surname></string-name>, <string-name>
<given-names>M.</given-names> 
<surname>Ba&#x010D;a</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>&#x0160;eva</surname></string-name>
</person-group>, &#x201C;
<article-title>Application of modified anthropometric model in facial age estimation</article-title>,&#x201D; in <conf-name>Proc. ELMAR-2011</conf-name>, 
<publisher-loc>Zadar, Croatia</publisher-loc>, pp. 
<fpage>17</fpage>&#x2013;
<lpage>20</lpage>, 
<year>2011</year>. </mixed-citation>
</ref>
<ref id="ref-49">
<label>49</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>J.</given-names> 
<surname>Jagtap</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Kokare</surname></string-name>
</person-group>, &#x201C;
<article-title>Human age classification using facial skin aging features and artificial neural network</article-title>,&#x201D; 
<source>Cognitive Systems Research</source>, vol. 
<volume>40</volume>, pp. 
<fpage>116</fpage>&#x2013;
<lpage>128</lpage>, 
<year>2016</year>.</mixed-citation>
</ref>
<ref id="ref-50">
<label>50</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S.</given-names> 
<surname>Bekhouche</surname></string-name>, <string-name>
<given-names>F.</given-names> 
<surname>Dornaika</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Benlamoudi</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Ouafi</surname></string-name> and <string-name>
<given-names>A.</given-names> 
<surname>Taleb-Ahmed</surname></string-name>
</person-group>, &#x201C;
<article-title>A comparative study of human facial age estimation: Handcrafted features vs. Deep features</article-title>,&#x201D; 
<source>Multimedia Tools and Applications</source>, vol. 
<volume>79</volume>, no. 
<issue>35</issue>, pp. 
<fpage>26605</fpage>&#x2013;
<lpage>26622</lpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-51">
<label>51</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S. E.</given-names> 
<surname>Choi</surname></string-name>, <string-name>
<given-names>Y. J.</given-names> 
<surname>Lee</surname></string-name>, <string-name>
<given-names>S. J.</given-names> 
<surname>Lee</surname></string-name>, <string-name>
<given-names>K. R.</given-names> 
<surname>Park</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>Kim</surname></string-name>
</person-group>, &#x201C;
<article-title>Age estimation using a hierarchical classifier based on global and local facial features</article-title>,&#x201D; 
<source>Pattern Recognition</source>, vol. 
<volume>44</volume>, no. 
<issue>6</issue>, pp. 
<fpage>1262</fpage>&#x2013;
<lpage>1281</lpage>, 
<year>2011</year>.</mixed-citation>
</ref>
<ref id="ref-52">
<label>52</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>Z.</given-names> 
<surname>Chai</surname></string-name>, <string-name>
<given-names>Z.</given-names> 
<surname>Sun</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Mendez-Vazquez</surname></string-name>, <string-name>
<given-names>R.</given-names> 
<surname>He</surname></string-name> and <string-name>
<given-names>T.</given-names> 
<surname>Tan</surname></string-name>
</person-group>, &#x201C;
<article-title>Gabor ordinal measures for face recognition</article-title>,&#x201D; 
<source>IEEE Transactions on Information Forensics and Security</source>, vol. 
<volume>9</volume>, no. 
<issue>1</issue>, pp. 
<fpage>14</fpage>&#x2013;
<lpage>26</lpage>, 
<year>2014</year>.</mixed-citation>
</ref>
<ref id="ref-53">
<label>53</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>M.</given-names> 
<surname>Haghighat</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Zonouz</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Abdel-Mottaleb</surname></string-name>
</person-group>, &#x201C;
<article-title>CloudID: Trustworthy cloud-based and cross-enterprise biometric identification</article-title>,&#x201D; 
<source>Expert Systems with Applications</source>, vol. 
<volume>42</volume>, no. 
<issue>21</issue>, pp. 
<fpage>7905</fpage>&#x2013;
<lpage>7916</lpage>, 
<year>2015</year>.</mixed-citation>
</ref>
<ref id="ref-54">
<label>54</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>A. Singh</given-names>, 
<surname>Sonal</surname></string-name> and <string-name>
<given-names>C.</given-names> 
<surname>Kant</surname></string-name>
</person-group>, &#x201C;
<article-title>Face and age recognition using three-dimensional discrete wavelet transform and rotational local binary pattern with radial basis function support vector machine method</article-title>,&#x201D; 
<source>International Journal of Electrical Engineering &#x0026; Education</source>, vol. 
<volume>58</volume>, no. 
<issue>2</issue>, pp. 
<fpage>0020720920988489</fpage>, 
<year>2021</year>.</mixed-citation>
</ref>
<ref id="ref-55">
<label>55</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>E.</given-names> 
<surname>Torres</surname></string-name>, <string-name>
<given-names>S. L.</given-names> 
<surname>Granizo</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Hernandez-Alvarez</surname></string-name>
</person-group>, &#x201C;
<article-title>Gender and age classification based on human features to detect illicit activity in suspicious sites</article-title>,&#x201D; in <conf-name>2019 Int. Conf. on Computational Science and Computational Intelligence</conf-name>, 
<publisher-loc>Las Vegas, NV, USA</publisher-loc>, pp. 
<fpage>416</fpage>&#x2013;
<lpage>419</lpage>, 
<year>2019</year>. </mixed-citation>
</ref>
<ref id="ref-56">
<label>56</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>P. C. S.</given-names> 
<surname>Reddy</surname></string-name>, <string-name>
<given-names>S. G.</given-names> 
<surname>Rao</surname></string-name>, <string-name>
<given-names>G.</given-names> 
<surname>Sakthidharan</surname></string-name> and <string-name>
<given-names>P. V.</given-names> 
<surname>Rao</surname></string-name>
</person-group>, &#x201C;
<article-title>Age grouping with central local binary pattern based structure co-occurrence features</article-title>,&#x201D; in <conf-name>2018 Int. Conf. on Smart Systems and Inventive Technology</conf-name>, 
<publisher-loc>Tirunelveli, India</publisher-loc>, pp. 
<fpage>104</fpage>&#x2013;
<lpage>109</lpage>, 
<year>2018</year>. </mixed-citation>
</ref>
<ref id="ref-57">
<label>57</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>M. E.</given-names> 
<surname>Malek</surname></string-name>, <string-name>
<given-names>Z.</given-names> 
<surname>Azimifar</surname></string-name> and <string-name>
<given-names>R.</given-names> 
<surname>Boostani</surname></string-name>
</person-group>, &#x201C;
<article-title>Facial age estimation using zernike moments and multi-layer perceptron</article-title>,&#x201D; in <conf-name>2017 22nd Int. Conf. on Digital Signal Processing</conf-name>, 
<publisher-loc>London, UK</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>5</lpage>, 
<year>2017</year>. </mixed-citation>
</ref>
<ref id="ref-58">
<label>58</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>X.-Q.</given-names> 
<surname>Zeng</surname></string-name>, <string-name>
<given-names>R.</given-names> 
<surname>Xiang</surname></string-name> and <string-name>
<given-names>H. X.</given-names> 
<surname>Zou</surname></string-name>
</person-group>, &#x201C;
<article-title>Partial least squares regression based facial age estimation</article-title>,&#x201D; in <conf-name>2017 IEEE Int. Conf. on Computational Science and Engineering and IEEE Int. Conf. on Embedded and Ubiquitous Computing</conf-name>, 
<publisher-loc>Guangzhou, China</publisher-loc>, pp. 
<fpage>416</fpage>&#x2013;
<lpage>421</lpage>, 
<year>2017</year>. </mixed-citation>
</ref>
<ref id="ref-59">
<label>59</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>R.</given-names> 
<surname>Angulu</surname></string-name>, <string-name>
<given-names>J. R.</given-names> 
<surname>Tapamo</surname></string-name> and <string-name>
<given-names>A. O.</given-names> 
<surname>Adewumi</surname></string-name>
</person-group>, &#x201C;
<article-title>Age estimation with local statistical biologically inspired features</article-title>,&#x201D; in <conf-name>2017 Int. Conf. on Computational Science and Computational Intelligence</conf-name>, 
<publisher-loc>Las Vegas, NV, USA</publisher-loc>, pp. 
<fpage>414</fpage>&#x2013;
<lpage>419</lpage>, 
<year>2017</year>. </mixed-citation>
</ref>
<ref id="ref-60">
<label>60</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>S.</given-names> 
<surname>Silvia</surname></string-name> and <string-name>
<given-names>S. H.</given-names> 
<surname>Supangkat</surname></string-name>
</person-group>, &#x201C;
<article-title>A literature review on facial age estimation researches</article-title>,&#x201D; in <conf-name>2020 Int. Conf. on ICT for Smart Society</conf-name>, 
<publisher-loc>Bandung, Indonesia</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>6</lpage>, 
<year>2020</year>. </mixed-citation>
</ref>
<ref id="ref-61">
<label>61</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>W.</given-names> 
<surname>Shen</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Guo</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Wang</surname></string-name>, <string-name>
<given-names>K.</given-names> 
<surname>Zhao</surname></string-name>, <string-name>
<given-names>B.</given-names> 
<surname>Wang</surname></string-name> <etal>et al.</etal>
</person-group><italic>,</italic> &#x201C;
<article-title>Deep differentiable random forests for age estimation</article-title>,&#x201D; 
<source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>, vol. 
<volume>43</volume>, no. 
<issue>2</issue>, pp. 
<fpage>404</fpage>&#x2013;
<lpage>419</lpage>, 
<year>2021</year>.</mixed-citation>
</ref>
<ref id="ref-62">
<label>62</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>H.</given-names> 
<surname>Zhang</surname></string-name>, <string-name>
<given-names>Y.</given-names> 
<surname>Zhang</surname></string-name> and <string-name>
<given-names>X.</given-names> 
<surname>Geng</surname></string-name>
</person-group>, &#x201C;
<article-title>Practical age estimation using deep label distribution learning</article-title>,&#x201D; 
<source>Frontiers of Computer Science</source>, vol. 
<volume>15</volume>, no. 
<issue>3</issue>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>6</lpage>, 
<year>2021</year>.</mixed-citation>
</ref>
<ref id="ref-63">
<label>63</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>Z.</given-names> 
<surname>Shah</surname></string-name> and <string-name>
<given-names>M.</given-names> 
<surname>Roy</surname></string-name>
</person-group>, &#x201C;
<article-title>Dcnn based dual assurance model for age estimation by analyzing facial images</article-title>,&#x201D; in <conf-name>2020 IEEE 17th India Council Int. Conf. (INDICON)</conf-name>, 
<publisher-loc>New Delhi, India</publisher-loc>, pp. 
<fpage>1</fpage>&#x2013;
<lpage>6</lpage>, 
<year>2020</year>. </mixed-citation>
</ref>
<ref id="ref-64">
<label>64</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>J. C.</given-names> 
<surname>Xie</surname></string-name> and <string-name>
<given-names>C. M.</given-names> 
<surname>Pun</surname></string-name>
</person-group>, &#x201C;
<article-title>Deep and ordinal ensemble learning for human age estimation from facial images</article-title>,&#x201D; 
<source>IEEE Transactions on Information Forensics and Security</source>, vol. 
<volume>15</volume>, pp. 
<fpage>2361</fpage>&#x2013;
<lpage>2374</lpage>, 
<year>2020</year>.</mixed-citation>
</ref>
<ref id="ref-65">
<label>65</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>H.-T. Q.</given-names> 
<surname>Bao</surname></string-name> and <string-name>
<given-names>S.-T.</given-names> 
<surname>Chung</surname></string-name>
</person-group>, &#x201C;
<article-title>A light-weight gender/age estimation model based on multi-taking deep learning for an embedded system</article-title>,&#x201D; in <conf-name>Proc. of the Korea Information Processing Society Conf.</conf-name>,&#x201D; in <conf-name>Korea Information Processing Society</conf-name>, pp. 
<fpage>483</fpage>&#x2013;
<lpage>486</lpage>, 
<year>2020</year>. </mixed-citation>
</ref>
<ref id="ref-66">
<label>66</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>H.</given-names> 
<surname>Pan</surname></string-name>, <string-name>
<given-names>H.</given-names> 
<surname>Han</surname></string-name>, <string-name>
<given-names>S.</given-names> 
<surname>Shan</surname></string-name> and <string-name>
<given-names>X.</given-names> 
<surname>Chen</surname></string-name>
</person-group>, &#x201C;
<article-title>Mean-variance loss for deep age estimation from a face</article-title>,&#x201D; in <conf-name>Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition</conf-name>, 
<publisher-loc>Salt Lake City, Utah</publisher-loc>, pp. 
<fpage>5285</fpage>&#x2013;
<lpage>5294</lpage>, 
<year>2018</year>. </mixed-citation>
</ref>
<ref id="ref-67">
<label>67</label><mixed-citation publication-type="journal">
<person-group person-group-type="author"><string-name>
<given-names>S.</given-names> 
<surname>Feng</surname></string-name>, <string-name>
<given-names>C.</given-names> 
<surname>Lang</surname></string-name>, <string-name>
<given-names>J.</given-names> 
<surname>Feng</surname></string-name>, <string-name>
<given-names>T.</given-names> 
<surname>Wang</surname></string-name> and <string-name>
<given-names>J.</given-names> 
<surname>Luo</surname></string-name>
</person-group>, &#x201C;
<article-title>Human facial age estimation by cost-sensitive label ranking and trace norm regularization</article-title>,&#x201D; 
<source>IEEE Transactions on Multimedia</source>, vol. 
<volume>19</volume>, no. 
<issue>1</issue>, pp. 
<fpage>136</fpage>&#x2013;
<lpage>148</lpage>, 
<year>2017</year>.</mixed-citation>
</ref>
<ref id="ref-68">
<label>68</label><mixed-citation publication-type="conf-proc">
<person-group person-group-type="author"><string-name>
<given-names>O.</given-names> 
<surname>Guehairia</surname></string-name>, <string-name>
<given-names>A.</given-names> 
<surname>Ouamane</surname></string-name>, <string-name>
<given-names>F.</given-names> 
<surname>Dornaika</surname></string-name> and <string-name>
<given-names>A.</given-names> 
<surname>Taleb-Ahmed</surname></string-name>
</person-group>, &#x201C;
<article-title>Deep random forest for facial age estimation based on face images</article-title>,&#x201D; in <conf-name>020 1st Int. Conf. on Communications, Control Systems and Signal Processing</conf-name>, 
<publisher-loc>El Oued, Algeria</publisher-loc>, pp. 
<fpage>305</fpage>&#x2013;
<lpage>309</lpage>, 
<year>2020</year>. </mixed-citation>
</ref>
</ref-list>
</back>
</article>