Research article

Quantitative analysis of facial proportions and facial attractiveness among Asians and Caucasians

  • Received: 20 November 2021 Revised: 26 February 2022 Accepted: 30 March 2022 Published: 22 April 2022
  • It has been proposed that the proportions of the human face are crucial for facial aesthetics. If this is the case, we should describe the relationship among proportions of face components quantitatively. This study aims to develop a mathematical model of facial proportions to provide a quantitative description of facial attractiveness. Furthermore, we expect that plastic surgeons can use models in clinical work to enhance communication efficiency between doctors and patients. Face alignment technique was used to analyse 5500 frontal faces with diverse properties (male/female, Asian/Caucasian, ages) to obtain the ratios among the nose length (NL), the nasal base width (N), and the inner canthus width (EI). A mathematical model (N2L=aEINL+bEIN+cNNL) was developed to describe the relationship among these proportions. To validate the effectiveness of this approach, we simulated the post-operative photos using Adobe Photoshop. Our findings show that the ratio of nose length to nose width, the ratio of inner canthus width to nose length and the ratio of inner canthus to nose width play a significant role in determining facial attractiveness. These results provide a possible strategy to quantitatively describe the relationship among human face proportions.

    Citation: Shikang Zheng, Kai Chen, Xinping Lin, Shiqian Liu, Jie Han, Guomin Wu. Quantitative analysis of facial proportions and facial attractiveness among Asians and Caucasians[J]. Mathematical Biosciences and Engineering, 2022, 19(6): 6379-6395. doi: 10.3934/mbe.2022299

    Related Papers:

    [1] Huiying Zhang, Jiayan Lin, Lan Zhou, Jiahui Shen, Wenshun Sheng . Facial age recognition based on deep manifold learning. Mathematical Biosciences and Engineering, 2024, 21(3): 4485-4500. doi: 10.3934/mbe.2024198
    [2] R Nandhini Abiram, P M Durai Raj Vincent . Identity preserving multi-pose facial expression recognition using fine tuned VGG on the latent space vector of generative adversarial network. Mathematical Biosciences and Engineering, 2021, 18(4): 3699-3717. doi: 10.3934/mbe.2021186
    [3] Qingwei Wang, Xiaolong Zhang, Xiaofeng Li . Facial feature point recognition method for human motion image using GNN. Mathematical Biosciences and Engineering, 2022, 19(4): 3803-3819. doi: 10.3934/mbe.2022175
    [4] Mubashir Ahmad, Saira, Omar Alfandi, Asad Masood Khattak, Syed Furqan Qadri, Iftikhar Ahmed Saeed, Salabat Khan, Bashir Hayat, Arshad Ahmad . Facial expression recognition using lightweight deep learning modeling. Mathematical Biosciences and Engineering, 2023, 20(5): 8208-8225. doi: 10.3934/mbe.2023357
    [5] Muhammad Ahmad Nawaz Ul Ghani, Kun She, Muhammad Arslan Rauf, Shumaila Khan, Masoud Alajmi, Yazeed Yasin Ghadi, Hend Khalid Alkahtani . Toward robust and privacy-enhanced facial recognition: A decentralized blockchain-based approach with GANs and deep learning. Mathematical Biosciences and Engineering, 2024, 21(3): 4165-4186. doi: 10.3934/mbe.2024184
    [6] Basem Assiri, Mohammad Alamgir Hossain . Face emotion recognition based on infrared thermal imagery by applying machine learning and parallelism. Mathematical Biosciences and Engineering, 2023, 20(1): 913-929. doi: 10.3934/mbe.2023042
    [7] Jianing Wang . Prediction of postoperative recovery in patients with acoustic neuroma using machine learning and SMOTE-ENN techniques. Mathematical Biosciences and Engineering, 2022, 19(10): 10407-10423. doi: 10.3934/mbe.2022487
    [8] Guangmin Sun, Hao Wang, Yu Bai, Kun Zheng, Yanjun Zhang, Xiaoyong Li, Jie Liu . PrivacyMask: Real-world privacy protection in face ID systems. Mathematical Biosciences and Engineering, 2023, 20(2): 1820-1840. doi: 10.3934/mbe.2023083
    [9] Qiming Li, Tongyue Tu . Large-pose facial makeup transfer based on generative adversarial network combined face alignment and face parsing. Mathematical Biosciences and Engineering, 2023, 20(1): 737-757. doi: 10.3934/mbe.2023034
    [10] C Willson Joseph, G. Jaspher Willsie Kathrine, Shanmuganathan Vimal, S Sumathi., Danilo Pelusi, Xiomara Patricia Blanco Valencia, Elena Verdú . Improved optimizer with deep learning model for emotion detection and classification. Mathematical Biosciences and Engineering, 2024, 21(7): 6631-6657. doi: 10.3934/mbe.2024290
  • It has been proposed that the proportions of the human face are crucial for facial aesthetics. If this is the case, we should describe the relationship among proportions of face components quantitatively. This study aims to develop a mathematical model of facial proportions to provide a quantitative description of facial attractiveness. Furthermore, we expect that plastic surgeons can use models in clinical work to enhance communication efficiency between doctors and patients. Face alignment technique was used to analyse 5500 frontal faces with diverse properties (male/female, Asian/Caucasian, ages) to obtain the ratios among the nose length (NL), the nasal base width (N), and the inner canthus width (EI). A mathematical model (N2L=aEINL+bEIN+cNNL) was developed to describe the relationship among these proportions. To validate the effectiveness of this approach, we simulated the post-operative photos using Adobe Photoshop. Our findings show that the ratio of nose length to nose width, the ratio of inner canthus width to nose length and the ratio of inner canthus to nose width play a significant role in determining facial attractiveness. These results provide a possible strategy to quantitatively describe the relationship among human face proportions.



    Abbreviations: AF: Asian females; AM: Asian males; CF: Caucasian females; CM: Caucasian males

    The face is vital in determining human physical beauty and attractiveness. Facial attractiveness is related to mate selection strategies, health, self-esteem and survival of species [1,2]. Attractive people experience more positive outcomes, such as, receiving attention and being more socially prevalent [3,4,5]. It has been proposed that facial appearance is an indicator of an individual's physical health [6].

    From ancient Greece to the Renaissance to modern society, people have been exploring facial attractiveness and the nature of beauty [7]. Many factors affect facial beauty and attractiveness, including: symmetry, youthfulness, skin texture, averageness, proportion, sexual dimorphism, good grooming, and even pregnancy [8,9,10,11,12,13].

    Proportions have a critical impact on facial attractiveness [14]. Coincidentally, both in China and the West, people have adopted the method of dividing the face horizontally into thirds and vertically into fifths to analyse and test a person's face [15]. The ideal ratio is related to the Golden Ratio, which dates back to the ancient Greek period and is widespread. It is represented by the letter Φ (phi, Φ ≈ 0.618) [16,17]. Marquardt applied the sacred ratio to three-dimensional space and designed the golden ratio mask, which has wide application [18]. However, this has been questioned by critics who consider this mask too masculine for females [19]. One method people can use to achieve optimal facial attractiveness is by changing proportion through plastic surgery [20]. Proportions are, therefore, an important component of facial analysis and thus essential for optimizing plastic surgery [21,22].

    Mateusz's et al. [23] results show that facial attractiveness increases as the uncovered eye surface increases and the nose and lip size decreases. However, only one variable was considered at a time, without examining the quantitative relationship between these variables. This raises the question of whether a mathematical formula can be established to describe the relationship among proportions of face components.

    In this study, we focused on the factors that could significantly affect a person's physical beauty with the help of face alignment techniques which allowed us to obtain 68 facial feature points. Furthermore, we tried to describe the relationship among proportions of face components in the mathematical formulation, determining the intrinsic links among these proportions.

    SCUT-FBP5500 is a new diverse benchmark dataset proposed by South China University of Technology (SCUT). It contains 5500 frontal faces in total with diverse properties (male/female, Asian/Caucasian, ages). The dataset includes 2000 Asian females, 2000 Asian males, 750 Caucasian females and 750 Caucasian males, along with beauty scores [1,5] and beauty score distributions. Scores were obtained by 60 volunteers aged from 18−27 (average 21.6) rating facial attractiveness on a 1−5 scale (Dataset was open on the website: https://github.com/HCIILAB/SCUT-FBP5500-Database-Release) [24].

    We selected clear photos of people without excessive make-up and occlusion from the dataset and, the face shown in the frontal aspect, with a neutral expression [25]. In total, 4740 photos were selected that met the criteria as the dataset for this experiment, including 1670 photos of Asia females (AF), 1778 photos of Asia males (AM), 627 photos of Caucasian females (CF), and 665 photos of Caucasian males (CM). The top 10% of scored photos make up the high attractiveness set, and the lowest-rated 10% of images make up the low attractiveness set.

    The photos were processed in Python using the "Dlib" and "OpenCV" Python packages.

    To reduce the arithmetic intensity, OpenCV was used to convert the images to grayscale (Figure 1b). The law of the gamma transformation, g(x,y)=(f(x,y))gamma, was used to remove non-uniform illumination of the images (Figure 1−c). Gamma values ranged from 0 to 1 and 0.3 was selected. Sixty-eight facial feature points of a face were extracted by the "Dlib" Python package (Figure 1d).

    Figure 1.  Image processing process.

    The 68 facial feature points contain essential information about a face, including the contour of the face, eyes, nose, mouth, etc. (Figure 2). The input to the procedure is a 2D image, and the output is a set of (x, y) coordinates of the facial landmark points. In this study, the inner canthus width (EI), nose length (NL), and nasal base width (N) were selected for facial analysis (Figure 3). The definition of the landmark [26] used in the analysis is shown in Table 1. The Euclidean distances of the above four facial feature quantities were calculated using:

    Inner canthus width:EI=(x39x42)2+(y39y42)2 (1)
    Length of Nose:NL=(x27x33)2+(y27y33)2 (2)
    Nasal base width:N=(x31x35)2+(y31y35)2 (3)
    Figure 2.  The distribution of 68 feature points model.
    Figure 3.  The quantities measured in this study.
    Table 1.  Definition of landmark used in analysis.
    27 Soft-tissue nasion 33 soft-tissue subnasal
    42 Left eye media canthus 39 Right eye media canthus
    35 Left Nasal ala 31 Right Nasal ala

     | Show Table
    DownLoad: CSV

    The ratios were then calculated between the quantities. R1 is the ratio of EI to N. R2 is the ratio of NL to N. R3 is the ratio of EI to NL.

    Independent samples t-test was used to analyse the differences in facial proportions(R1, R2 and R3) across gender and race. We compared the difference in facial proportions(R1, R2 and R3) between those with high attractiveness and those with low attractiveness. The normal distribution and variance homogeneity were tested by the P-P plot and Leven's test. In addition, a stepwise linear regression model was constructed to quantitively analyse the relationship among R1, R2 and R3.

    The data were processed using IBM SPSS Statistics 26 (IBM Corporation, Armonk, NY, USA) and P < 0.05 and P < 0.01 were acknowledged as statistically significant.

    A total of 20 images were selected from the low attractiveness set, including 5 images each from AF, AM, CF, CM. The nose length, inner canthus width, and nose width were obtained. An ideal inner canthus width was calculated by putting the nose length and the nose width into the formula derived from the high attractiveness set. The ideal nose width was obtained by the same method. Photoshop (Adobe Photoshop CC 2018) was used to separately change the inner canthus and the nasal base width in the photos to the ideal width. To vary the width of the inner canthus width and nose width, Filter-Liquify in Photoshop was used, allowing us to change the facial proportions quantitatively. Thus, three photographs were obtained for each image: the original, one with the nasal base width changed to the ideal, and one with the inner canthus width changed to the ideal. The photos were rated on a 1–5 scale by 48 volunteers ages 18−25 (average age 20.9, 24 females, 24 males), where 1 represented "extremely unattractive" and 5 represented "very attractive". Independent samples t-test was used to analyse the difference in attractiveness scores before and after the changes. P < 0.05 and P < 0.01 were acknowledged as statistically significant.

    The study was approved by the Ethics Committee of the Hospital of Stomatology, Jilin University (2020–38). Participants provided prior informed written consent.

    All procedures performed in studies involving human participants were in accordance with the ethical standards of the hospital research committee and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards.

    The description of facial proportions is presented in Table 2.

    Table 2.  Description of the facial proportions among AF、AM、CF and CM.
    Groups N Mean Std. Error Variance 95%CI Median
    AF R1 1670 1.4447 0.00249 0.010 1.4398~1.4496 1.4400
    R2 1670 1.8908 0.00342 0.020 1.8841~1.8975 1.8900
    R3 1670 0.7656 0.00120 0.002 0.7633~0.7680 0.7600
    AF
    Top 10%
    R1 167 1.4682 0.00839 0.012 1.4516~1.4848 1.4700
    R2 167 1.9766 0.01115 0.021 1.9546~1.9987 1.9700
    R3 167 0.7439 0.00281 0.001 0.7383~0.7494 0.7400
    AF
    Last 10%
    R1 167 1.4359 0.00733 0.009 1.4214~1.4503 1.4357
    R2 167 1.7998 0.01017 0.017 1.7797~1.8199 1.8100
    R3 167 0.8001 0.00451 0.003 0.7912~0.8090 0.8000
    AM R1 1778 1.4007 0.00242 0.010 1.3959~1.4054 1.4000
    R2 1778 1.8345 0.00328 0.019 1.8281~1.8410 1.8400
    R3 1778 0.7656 0.00128 0.003 0.7631~0.7681 0.7600
    AM
    Top 10%
    R1 178 1.4008 0.00765 0.010 1.3857~1.4159 1.4100
    R2 178 1.9395 0.00960 0.016 1.9206~1.9584 1.9300
    R3 178 0.7228 0.00302 0.002 0.7168~0.7287 0.7200
    AM
    Last 10%
    R1 178 1.3912 0.00751 0.010 1.3764~1.4061 1.3900
    R2 178 1.7521 0.01151 0.024 1.7294~1.7748 1.7500
    R3 178 0.7969 0.00435 0.003 0.7883~0.8055 0.8000
    CF R1 627 1.3927 0.00628 0.025 1.3804~1.4051 1.3800
    R2 627 1.8404 0.00789 0.039 1.8249~1.8559 1.8300
    R3 627 0.7583 0.00220 0.003 0.7540~0.7627 0.7500
    CF
    Top 10%
    R1 63 1.5510 0.01358 0.012 1.5238~1.5781 1.5500
    R2 63 2.0386 0.01589 0.016 2.0068~2.0703 2.0400
    R3 63 0.7616 0.00528 0.002 0.7510~0.7721 0.7600
    CF
    Last 10%
    R1 63 1.3170 0.01441 0.013 1.2882~1.3458 1.3200
    R2 63 1.6867 0.02129 0.029 1.6441~1.7292 1.6600
    R3 63 0.7848 0.00772 0.004 0.7693~0.8002 0.7800
    CM R1 665 1.2988 0.00511 0.017 1.2898~1.3098 1.2900
    R2 665 1.7835 0.00737 0.036 1.7690~1.7980 1.7800
    R3 665 0.7318 0.00234 0.004 0.7272~0.7364 0.7300
    CM
    Top 10%
    R1 67 1.4145 0.1503 0.015 1.3845~1.4445 1.4200
    R2 67 1.9987 0.1844 0.023 1.9618~2.0355 1.9700
    R3 67 0.7091 0.00609 0.002 0.6970~0.7213 0.7000
    CM
    Last 10%
    R1 67 1.2981 0.01581 0.017 1.2665~1.3296 1.2800
    R2 67 1.6778 0.02397 0.039 1.6299~1.7256 1.6500
    R3 67 0.7782 0.00863 0.005 0.7610~0.7954 0.7700

     | Show Table
    DownLoad: CSV

    Significant differences exist in facial proportions between race and gender. Independent samples t-test results showed that R1 was significantly higher in Asian females than in Asian males (t = 12.666, p < 0.01), Caucasian females (t = 7.685, p < 0.01) and Caucasian males (t = 25.471, p < 0.01). R2 was significantly higher in Asian females than in Asian males (t = 11.881, p < 0.01), Caucasian females (t = 5.865, p < 0.01) and Caucasian males (t = 13.204, p < 0.01). R3 was significantly lower in Caucasian males than in Asian females (t = 12.856, p < 0.01) and Asian males (t = 12.650, p < 0.01). Differences in facial proportions between race and gender are showed in Figure 4.

    Figure 4.  Differences of facial proportions between race and genders (*p < 0.05 vs AF, ●p < 0.05 vs AM Group, ⊙P < 0.05 vs CF Group).

    Significant differences exist in facial proportions between the high attractiveness set and low attractiveness set. Independent samples t-test results showed that images with high attractiveness scores were significantly higher in R1 than images with low attractiveness scores among Asian females (t = 2.901, p < 0.01), Caucasian females (t = 11.814, p < 0.01) and Caucasian males (t = 5.336, p < 0.01). Furthermore, images with high attractiveness scores were significantly higher in R2 than images with low attractiveness scores among Asian females (t = 11.719, p < 0.01), Asian males (t = 12.507, p < 0.01), Caucasian females (t = 13.245, p < 0.01) and Caucasian males (t = 10.610, p < 0.01). Additionally, images with high attractiveness scores were significantly lower in R3 than images with low attractiveness scores among Asian females (t = −10.579, p < 0.01), Asian males (t = −14.011, p < 0.01) Caucasian females (t = −2.477, p = 0.015) and Caucasian males (t = −6.545, p < 0.001). Differences of facial proportions between high attractiveness set and low attractiveness set are showed in Figure 5.

    Figure 5.  Differences of facial proportions between high attractiveness set and low attractiveness (*p < 0.05 vs All Group, ●p < 0.05 vs Top 10% Group).

    The mathematical model of R1, R2 and R3 among Asian females, Asian males, Caucasian females and Caucasian males was obtained after stepwise linear regression analysis: R2=aR1+bR3+c. The value of Durbin Watson was within the acceptable range, suggesting the independence of these variables. Furthermore, these variables presented no collinearity (VIF < 1.5). After a simple algebraic simplification, a polynomial equation with three variables was obtained: N2L=aEINL+bEIN+cNNL. The result of stepwise linear regression analysis among AF, AM, CF and CM and the result of stepwise linear regression analysis among high attractiveness and low attractiveness sets are presented in Table 3. The relationship between R1, R2 and R3 among AF, AM, CF and CM is presented in Figure 6.

    Table 3.  Results of stepwise linear regression analysis.
    Model B t Sig Tolerance VIF R2 Durbin-Watson
    AF Constant 1.886 322.14 0.000 0.991 1.943
    R1 1.300 372.29 0.000 0.842 1.188
    R3 −2.445 −335.87 0.000 0.842 1.188
    AF
    Top 10%
    Constant 1.999 92.81 0.000 0.992 1.735
    R1 1.325 137.64 0.000 0.883 1.132
    R3 −2.645 −90.44 0.000 0.883 1.132
    AF
    Last 10%
    Constant 1.838 103.75 0.000 0.990 2.116
    R1 1.230 −113.17 0.000 0.796 1.256
    R3 −2.255 99.76 0.000 0.796 1.256
    AM Constant 1.823 337.82 0.000 0.990 1.916
    R1 1.292 364.33 0.000 0.813 1.231
    R3 −2.349 −349.37 0.000 0.813 1.231
    AM
    Top 10%
    Constant 1.922 117.22 0.000 0.992 2.011
    R1 1.385 140.08 0.000 0.748 1.337
    R3 −2.661 −106.15 0.000 0.748 1.337
    AM
    Last 10%
    Constant 1.765 99.10 0.000 0.992 2.101
    R1 1.242 117.76 0.000 0.934 1.071
    R3 −2.184 −119.93 0.000 0.934 1.071
    CF Constant 1.833 179.04 0.000 0.992 2.058
    R1 1.307 265.13 0.000 0.851 1.176
    R3 −2.391 −170.00 0.000 0.851 1.176
    CF
    Top 10%
    Constant 2.082 76.22 0.000 0.992 1.493
    R1 1.313 83.17 0.000 0.713 1.403
    R3 −2.731 −67.29 0.000 0.713 1.403
    CF
    Last 10%
    Constant 1.665 55.01 0.000 0.991 2.373
    R1 1.334 71.71 0.000 0.896 1.116
    R3 −2.211 −63.69 0.000 0.896 1.116
    CM Constant 1.767 188.48 0.000 0.991 1.922
    R1 1.367 246.15 0.000 0.885 1.129
    R3 −2.405 −198.51 0.000 0.885 1.129
    CM
    Top 10%
    Constant 1.935 69.767 0.000 0.990 1.727
    R1 1.396 77.91 0.000 0.707 1.415
    R3 −2.696 −60.90 0.000 0.707 1.415
    CM
    Last 10%
    Constant 1.716 53.877 0.000 0.990 1.918
    R1 1.314 67.082 0.000 0.913 1.095
    R3 −2.241 −62.429 0.000 0.913 1.095
    Dependent Variable: R2; Predictors: (Constant), R1, R3

     | Show Table
    DownLoad: CSV
    Figure 6.  The relationship between R1, R2 and R3 among AF, AM, CF and CM.

    The models obtained by stepwise linear regression analysis for each group are as follows.

    AF:N2L=1.300EINL2.445EIN+1.886NNL(r2=0.991) (4)
    Top 10% AF:N2L=1.325NL2.645EIN+1.999NNL(r2=0.992) (5)
    Last 10% AF:N2L=1.230EINL2.255EIN+1.838NNL(r2=0.990) (6)
    AM:N2L=1.292EINL2.349EIN+1.823NNL(r2=0.990) (7)
    Top 10% AM:N2L=1.385EINL2.661EIN+1.922NNL(r2=0.992) (8)
    Last 10% AM:N2L=1.242EINL2.184EIN+1.765NNL(r2=0.992) (9)
    CF:N2L=1.307EINL2.391EIN+1.833NNL(r2=0.992) (10)
    Top 10% CF:N2L=1.313EINL2.731EIN+2.082NNL(r2=0.992) (11)
    Last 10% CF:N2L=1.334EINL2.211EIN+1.665NNL(r2=0.991) (12)
    CM:N2L=1.367EINL2.405EIN+1.767NNL(r2=0.991) (13)
    Top 10% CM:N2L=1.396EINL2.696EIN+1.935NNL(r2=0.990) (14)
    Last 10% CM:N2L=1.314EINL2.241EIN+1.716NNL(r2=0.990) (15)

    The mean of the original image scores for the five low attractiveness Asian females was 1.91. The mean of the scores after changing the width of the inner canthus was 2.40, significantly higher than the original image score (t = −7.306, p < 0.01). The mean of the scores after changing the width of the nose was 2.36, which was also significantly higher than the original image score (t = 6.779, p < 0.01). The mean of the original image scores for the five low attractiveness Asian males was 1.95. The mean of the scores after changing the width of the inner canthus was 2.53, significantly higher than the original image score (t = −8.424, p < 0.01). The mean of the scores after changing the width of the nose was 2.63, significantly higher than the original image score (t = 8.501, p < 0.01). The mean of the original image scores for the five low attractiveness Caucasian females was 1.86. The mean of the scores after changing the width of the inner canthus was 2.22, significantly higher than the original image score (t = −5.814, p < 0.01). The mean of the scores after changing the width of the nose was 2.18, significantly higher than the original image score (t = 4.982, p < 0.01). The mean of the original image scores for the five Caucasian males was 1.94. The mean score after changing the width of the inner canthus was 2.24, significantly higher than the original image score (t = −4.713, p < 0.01). The mean of the scores after changing the width of the nose was 1.97. The results of the independent samples t-test are shown in Figure 7.

    Figure 7.  Result of the verification (*p < 0.05 vs Origin Group, ●p < 0.05 vs Eye Group).

    In this study, we explored the relationship between facial proportions and attractiveness. We found that the ratio of inner canthus width to nose length, the ratio of nose length to nose width and the ratio of inner canthus width to nose length play a significant role in determining facial attractiveness in Asian females and Caucasian males. The ratio of nose length to nose width and the ratio of inner canthus width to nose length play a significant role in determining facial attractiveness in Asian males, whereas the ratio of inner canthus width to nose width and the ratio of nose length to nose width play an important role in determining facial beauty in Caucasian females. Furthermore, we developed a model to describe the relationship between the proportions of face components quantitatively.

    In contrast to Mateusz's et al. [23] findings, where the facial attractiveness increased as the nose size reduced, we found that the facial attractiveness increased as the ratio of NL to N increased among AF, AM, CF and CM. Additionally, the facial attractiveness increased as the ratio of EI to NL decreased among AF, AM and CM. Thus, a person's facial attractiveness increases when the length of the nose is longer, and the width of the nose is shorter. The size of the nose can be described in terms of length and width, and our experiment investigated the effect of these two quantities on facial attractiveness separately. Rather than discuss the effect of a particular proportion on facial attractiveness in isolation, we emphasized the relationship between facial proportions.

    No one aesthetic standard can represent universal beauty because people's perception of beauty is influenced by racial and cultural differences [15,27]. A common inaccuracy, which has existed since the 15th century, is that beauty can be defined and evaluated by proportions and linear parameters, dismissing racial and cultural differences [27,28,29,30,31]. For example, according to Da Vinci, the width of the mouth ideally equals the distance between the lips and the edge of the jaw [27,32]. Other findings suggest in a well-proportioned face the nasal base width should equal the inner canthus width [15]. In this experiment, ethnic and cultural differences in beauty perception were not considered, but it would be crucial to relate facial proportions to ethnic and cultural differences [27].

    The mathematical model (N2L=aEINL+bEIN+cNNL) was obtained after stepwise linear regression analysis. From different sets of facial pictures, we can obtain different parameters (a, b and c) allowing a person to select photos of faces based on their aesthetic preferences. Based on this set of photos, we can build a mathematical model about facial proportions. This method can thus consider the cultural-ethnic differences in facial analysis by taking into account the subjects' facial preferences.

    This mathematical model may provide a strategy for plastic surgical decision-making. In this mathematical model, three facial variables are involved: nose length, inner canthus width, and nasal base width. The length of the nose is usually difficult to alter by plastic surgery, whereas the surgery of the inner canthus and the nasal base width is relatively well established. Therefore, the patient's nose length can be taken as a fixed value. According to the patient's willingness and the plastic surgeon's evaluation, the inner canthus width or the nasal base width could be altered to achieve ideal facial proportions through plastic surgery. For example, if a patient wants to change the inner canthus width, we can measure the nasal base width as m and measure the nose length as n. The optimal inner canthus width can be calculated by simply bringing m and n into the formula. In this way, we can get the optimal inner canthus width: EI=n2cmnan+bm. Models can be built based on the patient's aesthetic preferences and used as a reference for plastic surgery. This ensures that plastic surgery outcomes are individualized, instead of one standard being applied to all.

    Good communication skills are essential for surgeons and for effective clinical practice [33,34]. Disputes between doctors and patients in plastic surgery are often reported [35]. This involves the issue of communication and harmonization of aesthetic standards between the patient and the physician in cosmetic surgery. Additionally, the modern biopsychosocial model of medicine requires a doctor to be a skilled operator of medical techniques and a person integrated with relevant knowledge and skills, which is more evident in aesthetic surgery [36,37,38]. Communication between surgeons and patients can be streamlined by allowing patients to select photos of faces based on his or her aesthetic preferences. A mathematical model can be produced based on this data, improving interactions between the practitioner and the patient, and allowing the patient's aesthetic preferences to be fully expressed.

    In recent years, various cutting-edge computer technologies, such as deep learning, have been used in facial analysis and other medical fields [39,40]. Through leveraging the marked attractive faces, Zhang et al. proposed a geometric beauty score function to evaluate attractiveness quantitively, modeled by the proposed semi-supervised HSSL learning method [41]. Zhang et al. quantitatively analysed the effect of facial geometric features on facial beauty, finding that the face can become more beautiful by making its geometric features closer to the average face shape [42]. In this experiment, we built a mathematical model from the SCUT-FBP5500, which reflects the average nature of facial features in the population. Verification tests have also demonstrated that the face can become more attractive as facial features are becoming closer to the average face shape. Xu et al. proposed a new network framework, classification and regression network (CRNet), to predict facial beauty [43]. Hong et al. presented a novel framework for automatically assessing facial attractiveness that considered four feature ratio sets as objective elements of facial attractiveness based on deep learning [44]. Tong's et al. findings show that a deep neural network (DNN) model can learn putative ratios from face images based only on categorical annotation when no annotated facial features for attractiveness are explicitly given [45]. He's et al. study shows that DiscoStyle can determine users' facial preferences reasoning and recommend preferred facial styles in different genders and races [46]. With the mathematical model obtained in this paper, we can further infer the aesthetic preferences of the users. Users can build a set of images based on their preferences that could then be used to create a mathematical model that could represent the user's preferences.

    Deep learning can predict facial attractiveness by establishing a nonlinear relationship among multiple features of the face. By using deep learning, the accuracy and efficiency of predicting facial attractiveness have been improved compared to traditional methods. Additionally, deep learning can consider a larger number of facial features.

    In this experiment, we focused on three facial features (EI,N and NL) that are frequently used in clinical practice. We developed linear regression models (N2L=aEINL+bEIN+cNNL) based on populations of different ethnicities and genders. Although we did not use deep learning in our experiments, the linear regression models all showed R2 values higher than 0.990, VIF < 1.5 and the value of Durbin Watson was within the acceptable range, indicating that the model establishes the relationship among the three facial features properly. Through validation tests, we also confirmed that facial attractiveness can be improved by changing the facial features of the participants with this model.

    There are four limitations to our study. First of all, in this experiment, we used 2D images, while the human brain processes 3D images [42]. Compared to 3D analysis, 2D images can result in the loss of facial information and lead to systematic errors. For example, the nose length (NL) in this experiment is not the actual length of the nose but the projected length of the actual nose length in the vertical plane where the face is located. Therefore, in future studies, it is recommended to use 3D images to study facial features. Secondly, the perception of attractiveness is affected by multiple other factors such as age [47] and female hormone status [48]. For example, female raters in different hormonal states will give different ratings to the same face and younger people have different aesthetic preferences from older people. Future studies should investigate the influence of these factors on mathematical models. Additionally, the perception of attractiveness is influenced by factors besides proportions, including symmetry, averageness, skin texture, and even filters [13,49,50,51]. Therefore, we encourage future studies to consider these factors. Thirdly, we did not try nonlinear modelling, such as neural networks. Only three facial feature quantities were selected in this study; other facial features should be investigated in the future. Last but not least, due to the aesthetic preferences of the East versus that of the West, the scores of 1500 Caucasians in the dataset rated by Chinese volunteers may not be representative of how Caucasians would rate the pictures.

    Our results show that the ratio of nose length to nose width, the ratio of inner canthus width to nose length and the ratio of inner canthus to nose width play a significant role in facial attractiveness among Asian females, Asian males, Caucasian females, and Caucasian males. We established a mathematical model (N2L=aEINL+bEIN+cNNL) to describe the relationship between a person's facial features. This model may provide some reference for plastic surgeons and facilitate better communication between doctors and patients. More facial feature quantities should be investigated in the future.

    This study was funded by Fund of Health Technology Innovation Project of Jilin Provincial Department of Finance (No.2019J031); Fund Project of Science and Technology Project of Jilin Provincial Department of Finance (No. JCSZ2019378-15); Fund Project of Science and Technology Project of Jilin Provincial Department of Finance (No. JCSZ2020304-10)

    Thanks to Jason Foster and Yang Guang for proofreading the article. We also thank He Yueyang and Meng Linghao of their recommendation for this mathematical model.

    The authors declare that there are no conflicts of interest regarding the publication of this paper.



    [1] D. W. Zaidel, S. M. Aarde, K. Baig, Appearance of symmetry, beauty, and health in human faces, Brain Cogn., 57 (2005), 261–263. https://doi.org/10.1016/j.bandc.2004.08.056 doi: 10.1016/j.bandc.2004.08.056
    [2] S. H. Mares, R. N. de Leeuw, R. H. Scholte, R. C. Engels, Facial attractiveness and self-esteem in adolescence, J. Clin. Child Adolesc. Psychol., 39 (2010), 627–637. https://doi.org/10.1080/15374416.2010.501292 doi: 10.1080/15374416.2010.501292
    [3] R. Adolphs, L. Nummenmaa, A. Todorov, J. V. Haxby, Data-driven approaches in the investigation of social perception, Philos. Trans. R. Soc. Lond. B Biol. Sci., 371 (2016), https://doi.org/10.1098/rstb.2015.0367 doi: 10.1098/rstb.2015.0367
    [4] B. Fink, N. Neave, The biology of facial beauty, Int. J. Cosmet. Sci., 27 (2005), 317–325. https://doi.org/10.1111/j.1467-2494.2005.00286.x doi: 10.1111/j.1467-2494.2005.00286.x
    [5] Y. Z. Foo, L. W. Simmons, G. Rhodes, Predictors of facial attractiveness and health in humans, Sci. Rep., 7 (2017), 39731. https://doi.org/10.1038/srep39731 doi: 10.1038/srep39731
    [6] A. Żelaźniewicz, J. Nowak, P. Łącka, B. Pawłowski, Facial appearance and metabolic health biomarkers in women, Sci. Rep., 10 (2020), 13067. https://doi.org/10.1038/s41598-020-70119-6 doi: 10.1038/s41598-020-70119-6
    [7] J. R. Thomas, More than a pretty face: the Mona Lisa, Arch. Facial Plast. Surg., 10 (2008), 65–66. https://doi.org/10.1001/archfaci.10.1.65 doi: 10.1001/archfaci.10.1.65
    [8] B. Fink, K. Grammer, R. Thornhill, Human (homo sapiens) facial attractiveness in relation to skin texture and color, J. Comp. Psychol., 115 (2001), 92–99. https://doi.org/10.1037/0735-7036.115.1.92 doi: 10.1037/0735-7036.115.1.92
    [9] J. Hong, B. Koo, J. Koo, The psychosocial and occupational impact of chronic skin disease, Dermatol. Ther., 21 (2008), 54–59. https://doi.org/10.1111/j.1529-8019.2008.00170.x doi: 10.1111/j.1529-8019.2008.00170.x
    [10] A. C. Little, B. C. Jones, C. Waitt, B. P. Tiddeman, D. R. Feinberg, D. I. Perrett, et al., Symmetry is related to sexual dimorphism in faces: data across culture and species, PLoS One, 3 (2008), e2106. https://doi.org/10.1371/journal.pone.0002106 doi: 10.1371/journal.pone.0002106
    [11] S. A. Milosević, M. L. Varga, M. Slaj, Analysis of the soft tissue facial profile of Croatians using of linear measurements, J. Craniofac. Surg., 19 (2008), 251–258. https://doi.org/10.1097/scs.0b013e31815c9446 doi: 10.1097/scs.0b013e31815c9446
    [12] N. Samson, B. Fink, P. J. Matts, Visible skin condition and perception of human facial appearance, Int. J. Cosmet. Sci., 32 (2010), 167–184. https://doi.org/10.1111/j.1468-2494.2009.00535.x doi: 10.1111/j.1468-2494.2009.00535.x
    [13] D. P. Danel, K. Kalinowski, N. Nowak-Szczepanska, A. Ziomkiewicz-Wichary, A. Apanasewicz, K. Borysławski, et al., Shifts in female facial attractiveness during pregnancy, Int. J. Environ. Res. Public Health, 17 (2020), 5176. https://doi.org/10.3390/ijerph17145176 doi: 10.3390/ijerph17145176
    [14] H. Shen, D. K. Chau, J. Su, L. L. Zeng, W. Jiang, J. He, et al., Brain responses to facial attractiveness induced by facial proportions: evidence from an fMRI study, Sci. Rep., 6 (2016), 35905. https://doi.org/10.1038/srep35905 doi: 10.1038/srep35905
    [15] H. Bueller, Ideal facial relationships and goals, Facial Plast. Surg., 34 (2018), 458–465. https://doi.org/10.1055/s-0038-1669401 doi: 10.1055/s-0038-1669401
    [16] Y. Jefferson, Facial beauty-establishing a universal standard, Int. J. Orthod. Milwaukee, 15 (2004), 9–22.
    [17] E. P. Prokopakis, I. M. Vlastos, V. A. Picavet, G. Nolst Trenite, R. Thomas, C. Cingi, et al., The golden ratio in facial symmetry, Rhinology., 51 (2013), 18–21. https://doi.org/10.4193/Rhino12.111 doi: 10.4193/Rhino12.111
    [18] S. R. Marquardt, Dr. Stephen R. Marquardt on the golden decagon and human facial beauty. interview by Dr. Gottlieb, J. Clin. Orthod., 36 (2002), 339–347.
    [19] E. Holland, Marquardt's Phi mask: pitfalls of relying on fashion models and the golden ratio to describe a beautiful face, Aesthetic Plast. Surg., 32 (2008), 200–208. https://doi.org/10.1007/s00266-007-9080-z doi: 10.1007/s00266-007-9080-z
    [20] K. S. Kaya, B. Türk, M. Cankaya, N. Seyhun, B. U. Coşkun, Assessment of facial analysis measurements by golden proportion, Braz. J. Otorhin., 85 (2019), 494–501. https://doi.org/10.1016/j.bjorl.2018.07.009 doi: 10.1016/j.bjorl.2018.07.009
    [21] M. Bashour, History and current concepts in the analysis of facial attractiveness, Plast. Reconstr. Surg., 118 (2006), 741–756. https://doi.org/10.1097/01.prs.0000233051.61512.65 doi: 10.1097/01.prs.0000233051.61512.65
    [22] D. Li, H. Xu, L. Xu, X. Huang, J. Yuan, M. Wei, The aesthetic proportion index of facial contour surgery, J. Craniofac. Surg., 26 (2015), 586–589. https://doi.org/10.1097/scs.0000000000001475 doi: 10.1097/scs.0000000000001475
    [23] M. Przylipiak, J. Przylipiak, R. Terlikowski, E. Lubowicka, L. Chrostek, A. Przylipiak, Impact of face proportions on face attractiveness, J. Cosmet. Dermatol., 17 (2018), 954–959. https://doi.org/10.1111/jocd.12783 doi: 10.1111/jocd.12783
    [24] L. Liang, L. Lin, L. Jin, D. Xie, M. Li, SCUT-FBP5500: a diverse benchmark dataset for multi-paradigm facial beauty prediction, in 2018 24th International Conference on Pattern Recognition (ICPR), (2018), 1598–1603. https://doi.org/10.1109/ICPR.2018.8546038
    [25] M. Matyjek, M. Bayer, I. Dziobek, Pupillary responses to faces are modulated by familiarity and rewarding context, Brain Sci., 11 (2021). https://doi.org/10.3390/brainsci11060794 doi: 10.3390/brainsci11060794
    [26] V. F. Ferrario, C. Sforza, G. Pizzini, G. Vogel, A. Miani, Sexual dimorphism in the human face assessed by euclidean distance matrix analysis, J. Anat., 183 (1993), 593–600.
    [27] A. Sisti, N. Aryan, P. Sadeghi, What is beauty?, Aes. Plast Surg., 45 (2021), 2163–2176. https://doi.org/10.1007/s00266-021-02288-2 doi: 10.1007/s00266-021-02288-2
    [28] T. Greywal, S. H. Dayan, K. Goldie, S. Guillen Fabi, The perception bias of aesthetic providers, J. Cosmet. Dermatol., 20 (2021), 1618–1621. https://doi.org/10.1111/jocd.13785 doi: 10.1111/jocd.13785
    [29] P. N. Broer, S. Juran, Y. J. Liu, K. Weichman, N. Tanna, M. E. Walker, et al., The impact of geographic, ethnic, and demographic dynamics on the perception of beauty, J. Craniofac. Surg., 25 (2014), e157–161. https://doi.org/10.1097/scs.0000000000000406 doi: 10.1097/scs.0000000000000406
    [30] A. Sisti, P. Sadeghi, Leonardo da Vinci: the Lady with an Ermine enigma, Hormones, 20 (2020), 591–592. https://doi.org/10.1007/s42000-020-00264-0 doi: 10.1007/s42000-020-00264-0
    [31] A. C. Little, Facial attractiveness, Wiley Inter. Rev. Cogn. Sci., 5 (2014), 621–634. https://doi.org/10.1002/wcs.1316 doi: 10.1002/wcs.1316
    [32] F. Vegter, J. J. Hage, Clinical anthropometry and canons of the face in historical perspective, Plast Reconstr. Surg., 106 (2000), 1090–1096. https://doi.org/10.1097/00006534-200010000-00021 doi: 10.1097/00006534-200010000-00021
    [33] S. Nakagawa, K. Fischkoff, A. Berlin, T. D. Arnell, C. D. Blinderman, Communication skills training for general surgery residents, J. Surg. Educ., 76 (2019), 1223–1230. https://doi.org/10.1016/j.jsurg.2019.04.001 doi: 10.1016/j.jsurg.2019.04.001
    [34] S. M. Weldon, T. Korkiakangas, J. Bezemer, R. Kneebone, Communication in the operating theatre, Br. J. Surg., 100 (2013), 1677–1688. https://doi.org/10.1002/bjs.9332 doi: 10.1002/bjs.9332
    [35] M. Singh, Communication as a bridge to build a sound doctor-patient/parent relationship, Indian J. Pediatr., 83 (2016), 33–37. https://doi.org/10.1007/s12098-015-1853-9 doi: 10.1007/s12098-015-1853-9
    [36] P. G. Morselli, A. Lippi, F. A. Giorgini, E. Fabbri, V. Pinto, Informed consent in plastic surgery, evaluation of its effectiveness for mutual satisfaction of patient and doctor: comparison of methods, J. Plast Reconstr. Aesthet. Surg., 72 (2019), 1847–1855. https://doi.org/10.1016/j.bjps.2019.05.037 doi: 10.1016/j.bjps.2019.05.037
    [37] G. L. Engel, The clinical application of the biopsychosocial model, Am. J. Psychiatry, 137 (1980), 535–544. https://doi.org/10.1176/ajp.137.5.535 doi: 10.1176/ajp.137.5.535
    [38] Y. Alonso, The biopsychosocial model in medical research: the evolution of the health concept over the last two decades, Patient Educ. Couns., 53 (2004), 239–244. https://doi.org/10.1016/s0738-3991(03)00146-0 doi: 10.1016/s0738-3991(03)00146-0
    [39] R. Patcas, R. Timofte, A. Volokitin, E. Agustsson, T. Eliades, M. Eichenberger, et al., Facial attractiveness of cleft patients: a direct comparison between artificial-intelligence-based scoring and conventional rater groups, Eur. J. Orthod., 41 (2019), 428–433. https://doi.org/10.1093/ejo/cjz007 doi: 10.1093/ejo/cjz007
    [40] M. Nolden, S. Zelzer, A. Seitel, D. Wald, M. Müller, A. M. Franz, et al., The medical imaging interaction toolkit: challenges and advances : 10 years of open-source development, Int. J. Comput. Assist. Radiol. Surg., 8 (2013), 607–620. https://doi.org/10.1007/s11548-013-0840-8 doi: 10.1007/s11548-013-0840-8
    [41] L. Zhang, D. Zhang, M. M. Sun, F. M. Chen, Facial beauty analysis based on geometric feature: toward attractiveness assessment application, Expert Syst. Appl., 82 (2017), 252–265. https://doi.org/10.1016/j.eswa.2017.04.021 doi: 10.1016/j.eswa.2017.04.021
    [42] D. Zhang, Q. Zhao, F. Chen, Quantitative analysis of human facial beauty using geometric features, Pattern Recogn., 44 (2011), 940–950. https://doi.org/10.1016/j.patcog.2010.10.013 doi: 10.1016/j.patcog.2010.10.013
    [43] L. Xu, J. Xiang, X. Yuan, CRNet: Classification and regression neural network for facial beauty prediction, 11166 (2018), 661–671. https://doi.org/10.1007/978-3-030-00764-5_61
    [44] Y. J. Hong, G. P. Nam, H. Choi, J. Cho, I. J. Kim, A novel framework for assessing facial attractiveness based on facial proportions, Symmetry (Basel), 9 (2017), 294. https://doi.org/10.3390/sym9120294 doi: 10.3390/sym9120294
    [45] S. Tong, X. Liang, T. Kumada, S. Iwaki, Putative ratios of facial attractiveness in a deep neural network, Vision Res., 178 (2021), 86–99. https://doi.org/10.1016/j.visres.2020.10.001 doi: 10.1016/j.visres.2020.10.001
    [46] Z.-W. He, L. Zhang, F. Y. Liu, DiscoStyle: multi-level logistic ranking for personalized image style preference inference, Int. J. Autom. Comput., 17 (2020), 637–651. https://doi.org/10.1007/s11633-020-1244-1 doi: 10.1007/s11633-020-1244-1
    [47] C. Han, X. Lei, P. Yan, X. Li, E. R. Morrison, Age differences in preferences for body physique, Pers. Indiv. differ., 181 (2021), 111033. https://doi.org/10.1016/j.paid.2021.111033 doi: 10.1016/j.paid.2021.111033
    [48] K. Gildersleeve, M. G. Haselton, M. R. Fales, Do women's mate preferences change across the ovulatory cycle? a meta-analytic review, Psychol. Bull., 140 (2014), 1205–1259. https://doi.org/10.1037/a0035438 doi: 10.1037/a0035438
    [49] A. C. Little, B. C. Jones, L. M. DeBruine, Facial attractiveness: evolutionary based research, Philos. Trans. Roy. Soc., 366 (2011), 1638–1659. https://doi.org/10.1098/rstb.2010.0404 doi: 10.1098/rstb.2010.0404
    [50] M. Komori, S. Kawamura, S. Ishihara, Averageness or symmetry: which is more important for facial attractiveness?, Acta Psychol., 131 (2009), 136–142. https://doi.org/10.1016/j.actpsy.2009.03.008 doi: 10.1016/j.actpsy.2009.03.008
    [51] G. Vera Cruz, The impact of face skin tone vs. face symmetry on perceived facial attractiveness, J. Gen. Psychol., 145 (2018), 183–198. https://doi.org/10.1080/00221309.2018.1459452 doi: 10.1080/00221309.2018.1459452
  • This article has been cited by:

    1. Jihyeon Yeom, Jeongin Lee, Heekyung Yang, Kyungha Min, Perception of Facial Impressions Using Explicit Features of the Face (xFoFs), 2023, 11, 2227-7390, 3779, 10.3390/math11173779
    2. Randa AlFotawi, AlJoharah AlShathry, Hafsah AlAli, Ibrahim AlOmran, Njood AlSayari, Reema AlShahrani, Ahmad Alzahrani, The Perception Facial Attractiveness (Pfa) in Saudi Culture Through the Lens of the Golden Ratio, 2025, 0972-8279, 10.1007/s12663-025-02477-5
    3. Randa ALFOTAWI, Aljoharah ALSHATHRY, Hafsah AL ALI, Ibrahim ALOMRAN, Njood ALSAYARI, Reema ALSHAHRANI, Ahmad ALZAHRANI, The perception facial attractiveness (PFA) in Saudi culture through the lens of the golden ratio, 2025, 38, 03949508, 10.23736/S0394-9508.24.05813-3
    4. Zhi-Lin Chen, Kang-Ming Chang, Exploring the Correlation Between Gaze Patterns and Facial Geometric Parameters: A Cross-Cultural Comparison Between Real and Animated Faces, 2025, 17, 2073-8994, 528, 10.3390/sym17040528
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(4603) PDF downloads(295) Cited by(4)

Figures and Tables

Figures(7)  /  Tables(3)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog