
Citation: Jinhua Zeng, Xiulian Qiu, Shaopei Shi. Image processing effects on the deep face recognition system[J]. Mathematical Biosciences and Engineering, 2021, 18(2): 1187-1200. doi: 10.3934/mbe.2021064
[1] | Hao Wang, Guangmin Sun, Kun Zheng, Hui Li, Jie Liu, Yu Bai . Privacy protection generalization with adversarial fusion. Mathematical Biosciences and Engineering, 2022, 19(7): 7314-7336. doi: 10.3934/mbe.2022345 |
[2] | Basem Assiri, Mohammad Alamgir Hossain . Face emotion recognition based on infrared thermal imagery by applying machine learning and parallelism. Mathematical Biosciences and Engineering, 2023, 20(1): 913-929. doi: 10.3934/mbe.2023042 |
[3] | Shuai Cao, Biao Song . Visual attentional-driven deep learning method for flower recognition. Mathematical Biosciences and Engineering, 2021, 18(3): 1981-1991. doi: 10.3934/mbe.2021103 |
[4] | Shangbin Li, Yu Liu . Human motion recognition based on Nano-CMOS Image sensor. Mathematical Biosciences and Engineering, 2023, 20(6): 10135-10152. doi: 10.3934/mbe.2023444 |
[5] | Qingwei Wang, Xiaolong Zhang, Xiaofeng Li . Facial feature point recognition method for human motion image using GNN. Mathematical Biosciences and Engineering, 2022, 19(4): 3803-3819. doi: 10.3934/mbe.2022175 |
[6] | Xiao Ma, Xuemei Luo . Finger vein recognition method based on ant colony optimization and improved EfficientNetV2. Mathematical Biosciences and Engineering, 2023, 20(6): 11081-11100. doi: 10.3934/mbe.2023490 |
[7] | Zilong Liu, Jingbing Li, Jing Liu . Encrypted face recognition algorithm based on Ridgelet-DCT transform and THM chaos. Mathematical Biosciences and Engineering, 2022, 19(2): 1373-1387. doi: 10.3934/mbe.2022063 |
[8] | Qiaokang Liang, Jianzhong Peng, Zhengwei Li, Daqi Xie, Wei Sun, Yaonan Wang, Dan Zhang . Robust table recognition for printed document images. Mathematical Biosciences and Engineering, 2020, 17(4): 3203-3223. doi: 10.3934/mbe.2020182 |
[9] | Qi Cui, Ruohan Meng, Zhili Zhou, Xingming Sun, Kaiwen Zhu . An anti-forensic scheme on computer graphic images and natural images using generative adversarial networks. Mathematical Biosciences and Engineering, 2019, 16(5): 4923-4935. doi: 10.3934/mbe.2019248 |
[10] | Qian Zhang, Haigang Li, Ming Li, Lei Ding . Feature extraction of face image based on LBP and 2-D Gabor wavelet transform. Mathematical Biosciences and Engineering, 2020, 17(2): 1578-1592. doi: 10.3934/mbe.2020082 |
Currently, face recognition systems (FRS) [1,2] have play an important role in face recognition and face verification tasks. In the field of public security, we can use face images of perpetrators at crime scenes to obtain possible identification of suspects by searching and comparison through the FRS. And in the subsequent examination of forensic identification of human images, judicial experts can give professional judgment on whether the identity of the perpetrator and the suspect are the same. After that, the image materials can be better converted into court evidence to demonstrate the facts of crimes.
Forensic identification of human images (FIHI) is to make the professional judgment on whether human images captured in images or videos are from the same persons [3]. Usually, we name the human images in crime scenes as the questioned human images, and the suspect images as the known human images. In the process of giving expert opinions, the experts use both the qualitative and quantitative analysis methods. Currently, the face recognition technology is one of the most important quantitative examination methods for FIHI, which can carry out efficient and accurate quantitative inspection and analysis on face features of facial components and their spatial relationships [4]. The human image features in the field of FIHI consist of the head morphological features, facial components and their spatial relationship features, beard features, wrinkle features, facial dynamic features, body type features, special human body marking features, human dress features, body adornment features, and other human dynamic characteristics, such as gait [5]. Among them, the facial components and their spatial relationship features are one of the most important parts of the human image features, and they are usually the key features in most cases of FIHI.
However, the real problem is that in practical cases, the questioned face images, for example, captured in road monitoring systems are usually in poor quality due to long shooting distances or low image resolutions, which may have a great impact on the recognition performance of FRS [6]. What's more, in many cases, it even does not meet the minimum input requirement of FRS.
The field of image processing research has a long history. Wealth of image processing techniques have been proposed. Image processing methods can be roughly divided into image denoising [7,8] and image enhancement techniques [9] according to the different usage purposes. Image denoising is mainly to remove image information that is useless or interfering with the target information, including from the traditional filter-based denoising methods [10] to current deep denoising techniques [11,12].
Image denoising is a classic problem and has been studied for a long time. So far, researchers have proposed various methods for image denoising. The state-of-the-art single image denoising techniques can be divided into filtering, sparse coding, classical external priors, low rank, deep learning and hybrid model based methods [13,14,15,16]. In addition, image denoising methods can be generally classified into spatial domain methods and transform domain methods [17]. Spatial domain based image denoising, for example Wiener filtering [18], relies on an assumption that noises occupy the higher region of the frequency spectrum, and they make use of low pass filtering on pixel groups. Its disadvantage is the negative effect on image blurring, which in turn loses sharp edges [10]. In the transform domain based image denoising, it first transforms images to specific domain, such as the discrete cosine transform domain, and then different denoising strategies are applied in transformed images with the statement that noise features are different in the transform domain. For example, the wavelet transform [19] decomposes images into scale-space representation, and then removes noises while preserving image features regardless of frequency content. The 3DDCT based image denoising [20] was to enlarge non-zero DCT coefficients in high frequency parts which would enhance the image texture and edge information while denoising.
The image enhancement techniques are mainly used to highlight useful information in images, such as the edge-preserving self-snake model based methods [21]. These technologies can more or less improve people's visual perception experiences of images. Additional, some methods were proposed to specialize in face image super-resolution [22,23]. Among them, the learning-based image super-resolution techniques [24,25,26] are one of the hotest solutions for increasing the spatial resolution of low-resolution images. Especially, recent years have witnessed remarkable progress of deep learning based image super-resolution methods. However, in the field of forensic science, it requires the submitted data with the features of the authenticity, relevance, and legality. In the field of FIHI, more emphasis is placed on the authenticity of data sources. Therefore, the image processing should not change substantive contents of original face images. This requires us to hold a cautious attitude in face image processing with regard to the additional data introduced in image processing.
Several literatures [27,28,29] existed on face image quality assessment via deep convolutional neural network based face recognition techniques. The studied quality factors of face images comprised the illumination [27], contrast, blurriness, occlusion, pose [28], low-resolution, noise [29] etc. For the research of the effects of image quality or image processing methods on the recognition performance of FRS, the effects of the covariate of lighting conditions on the recognition precision of face recognition algorithms was studied in [30], in which the results showed that poor lighting conditions dramatically decreased the face recognition performance. Bharadwaj et al. [6] studied the effects of the BayesShrink denoising techniques on the local binary patterns [31] based texture recognition method for face recognition. The BayesShrink technique [32] was wavelet based soft thresholding technique for image denoising, and the experiments were carried out by using the AR face database [33]. The influence of different denoising methods on the recognition performance of seven holistic 3D face recognition algorithms was studied in [34], in which six denoising algorithms were evaluated which consisted of Gaussian, mean and median filtering, multi-scale wavelet denoising, adaptive Wiener filtering and non-linear diffusion. And the seven 3D face recognition algorithms were studied, which comprised multi-class support vector machine, principal component analysis, kernel Fisher's analysis, probabilistic neural network, KNN-classification, bootstrap agammaegation decision trees and linear discriminant analysis methods. The experiments were conducted by using the Face Recognition Grand Challenge v2.0 dataset [35].
The existing research on the face image quality assessment methods [28,29] was aimed to select optimal face images with high quality for face recognition in FRS, and considered the image quality factors comprised noises and others. However, they were very different with our work, in which we wanted to study the effects of the image processing on the recognition performance of the deep face recognition system. In addition, in our studied topic, similar literature [6,34] focused on the FRS with traditional face recognition algorithms, such as the local binary patterns and support vector machine, which had substantial performance differences compared with currently widely used deep learning based face recognition techniques [37,38]. Furthermore, their studied noises were manually synthesized and added in the face images [6,29] rather than using noises under natural environmental corruption. The state-of-the-art face recognition technology are typically based on convolutional neural network based methods [37,38], on which the effects of varieties of image processing techniques are still unknown, especially in the field of real case studies. We aimed to study the effects of image denoising and enhancement techniques on the deep face recognition system under the application of FIHI. In order to enhance the validity and practical applicability of the results, the face image materials were all derived from real cases in FIHI. We collected 33 real cases which all happened in recent years. In these cases, the initial information of the suspects was found out through the recognition of the perpetrators' face images in FRS. And in the subsequent examination of FIHI, the authors served as the judicial experts giving the expert opinions that the perpetrator and the suspect were the same persons in each cases. Furthermore, the above expert opinions were all accepted by courts. The practical case study could guarantee the research results more suitable for practical application in FIHI. Our contribution could be summarized as follows:
(1) Understanding the effect of the image processing techniques on the state-of-the-art neural network based FRS under the real cases of FIHI, including the image denoising and enhancement techniques.
(2) Quantitative analysis of the effect of the image processing methods on the FRS. Investigating how many effects of the image processing techniques can bring on the recognition performance of FRS.
(3) Statistical comparative analysis of the effects of the Gaussian filtering based image denoising and the self-snake model based image enhancement techniques on the face recognition system.
The rest of the paper was organized as follows: Section 2 comprised the basic information of the tested face image materials, including their source, image quality, etc. The detailed experimental methods also presented in this section. The experimental results were shown in section 3, in which the statistical effects of the tested image quality and the image processing methods were described. The discussions and conclusions were given in the last section.
In order to guarantee the generalization and applicability of the results in practical case applications, the used face image materials were all from real cases in the field of forensic identification of human images. The state-of-the-art deep face recognitin system was used and tested to study the effects of different image processing techniques. The Gaussian filtering based image denoising method and the self-snake model based image enhancement method was studied which were commonly used in the field. In order to evaluate the effects of the image processing methods, the original face images without processing were used as a control point, and were compared with the processed ones. The processing operation was only applied to the questioned face images and the known face images were directly used without image processing. Their effects were measured by the recognition precision in the face verification tasks of the studied FRS. The main steps of our study were shown in Figure 1.
Thirty three pairs of the questioned face images and the known face images were collected from the 33 real cases of the FIHI. The size distribution of the questioned face images captured by the face detection method was shown in Figure 2, in which the smallest face image size was 30*28 pixels and the largest one was 175*174 pixels. The known face images were from the photos of the suspects' identity card or the front faces of the suspects taken when they were captured, which generally had better image quality. We studied the effect of the image processing methods applied to the questioned face images on the state-of-the-art neural network based FRS. The factor of the known face images processing was not considered in this paper.
We used the face recognition system which was based on the MXNet system architecture to quantify the similarity between the 33 pairs of the questioned and the known face images. In the used FRS, the scores of the face verification task were between 0 and 100. The higher score meant the more similar between the compared face images. In order to study the effect of the image processing techniques. The similarity values were computed between the known face images and the questioned face images before and after image processing respectively. The statistical data analysis techniques were used to compute the effect of the image processing methods on the FRS.
The simplest Gaussian filtering method [10] was selected to denoise face images, and the self-snake based method [39] was used to enhance face images which was a most basic and important image enhancement method based on Partial Differential Equation [39,40]. The selected two image processing methods were very common and popular in the field of the FIHI. In the Gaussian filtering based image denoising, the parameters of the Gaussian blur radius were set to 0.5, 1.0 and 2.0 pixels respectively. We set the zoom ratio of the face images to 200% in the self-snake based face image enhancement. We used the JC(i)(i = 1, ..., 33) and YB(i)(i = 1, ..., 33) to represent the sets of the questioned face images and the known face images in the cases. GB0.5(i)(i = 1, ..., 33), GB1(i)(i = 1, ..., 33), GB2(i)(i = 1, ..., 33) were used to represent the JC(i)(i = 1, ..., 33) denoised by the Gaussian filtering with blur radii of 0.5, 1.0, and 2.0 pixels respectively. SS(i)(i = 1, ..., 33) meant JC(i)(i = 1, ..., 33) processed by the self-snake model based image enhancement method. The similarity scores between JC(i)(i = 1, ..., 33) and YB(i)(i = 1, ..., 33) calculated in the FRS were marked as SM(i)(i = 1, ..., 33). The SM1(i)(i = 1, ..., 33), SM2(i)(i = 1, ..., 33), SM3(i)(i = 1, ..., 33), SMs(i)(i = 1, ..., 33) respectively were used to represent the similarity scores of the GB0.5(i)(i = 1, ..., 33), GB1(i)(i = 1, ..., 33), GB2(i)(i = 1, ..., 33), and SS(i)(i = 1, ..., 33) compared with YB(i)(i = 1, ..., 33) in the face verification tasks.
Considering the experimental design methods, the repeated measurement analysis of variance method was used to analyze the effect of different image processing methods on the FRS. The influence of the face image size on the recognition performance of the FRS was tested by means of one-way analysis of variance.
In order to facilitate data analysis, we divided the face image sizes into three categories, ie, the width and height of the face images were less than 50 pixels, between 50 and 70 pixels, and more than 70 pixels, marked as Size1, Size2, and Size3. The one-way ANOVA analysis was carried out to analyze the effect of the face image size on the FRS. The results were shown in Table 1.
SizeNo | Mean | Std. Deviation | N |
Size1 | 80.142 | 4.8286 | 12 |
Size2 | 84.255 | 4.7798 | 11 |
Size3 | 86.440 | 6.1661 | 10 |
Total | 83.421 | 5.7417 | 33 |
The homogeneity test of variance showed that the variance was homogeneous (F(2, 30) = 0.812, P = 0.453). Our results showed that there were significant differences in the means (F(2, 30) = 4.132, P = 0.026). The pairwise comparison test among the image sizes found that compared with the Size 1 face images, the Size 3 obtained higher face verification scores in the FRS (F = 6.298, P = 0.026). The scores of the Size 2 were not significantly different from the Size 1 (P = 0.070) and Size 3 (P = 0.348). The results showed that the bigger sizes of face images, the higher scores the face verification results could be obtained. Generally, the larger face image sizes meant the higher image quality which then could obtain the higher scores in verification tasks of the FRS.
In order to study the effects of the different image processing methods on the recognition performance of the FRS, we selected the highest scores of the Gaussian filtering based methods with different parameter setting as the final effects of the denoising method, labelled as SMg(i)(i = 1, ..., 33). Namely, SMg(i)(i = 1, ..., 33) = Max{SM1(i), SM2(i), SM3(i)}(i = 1, ..., 33). We used the repeated measurement analysis of variance method to compare the mean differences among the SM(i)(i = 1, ..., 33), SMg(i)(i = 1, ..., 33) and SMs(i)(i = 1, ..., 33). The descriptive statistics information were shown in Figure 3.
The Greenhouse-Geisser method (P = 0.797) was used to analyze the tested data since they did not meet the Mauchly's Test of Sphericity (P = 0.010). The results showed that there were significant differences among the SM(i)(i = 1, ..., 33), SMg(i)(i = 1, ..., 33) and SMs(i)(i = 1, ..., 33) (F(1.593, 50.977) = 40.986, P = 0.000). The subsequent pairwise comparison examinations were further carried out, and the results were described as follows.
3.2.1.SM(i)(i = 1, ..., 33) vs. SMg(i)(i = 1, ..., 33) and SMs(i)(i = 1, ..., 33)
The comparison results between SM(i)(i = 1, ..., 33) vs. SMg(i)(i = 1, ..., 33) and SMs(i)(i = 1, ..., 33) were shown in Table 2.
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SM | SMg | −2.445 * | 0.309 | 0.000 | −3.226 | −1.665 |
SMs | −0.864 * | 0.193 | 0.000 | −1.351 | −0.377 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
The results showed that there was significant difference between SM(i)(i = 1, ..., 33) and SMg(i)(i = 1, ..., 33), which meant that comparing to the original face images, the face images with the Gaussian filtering based image denoising could obtain higher similarity scores in the FRS. The effect of the Gaussian filtering based image denoising was 2.445. Although the Gaussian filtering processing reduced the image sharpness, it improved the recognition performance of the FRS. There was significant difference between SM(i)(i = 1, ..., 33) and SMs(i)(i = 1, ..., 33), that was, comparing to the original face images, the face images with the self-snake based image enhancement could improve the recognition performance of the FRS, and their mean differences was 0.864. The results showed that the self-snake model based image enhancement not only enhanced the visual display result of images, but it also improved the recognition precision of the FRS.
3.2.2.SMg(i)(i = 1, ..., 33) vs. SMs(i)(i = 1, ..., 33)
The comparison results between SMg(i)(i = 1, ..., 33) and SMs(i)(i = 1, ..., 33) were shown in Table 3.
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SMg | SM | −2.445 * | 0.309 | 0.000 | −3.226 | −1.665 |
SMs | 1.582 * | 0.304 | 0.000 | 0.813 | 2.350 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
The results showed that there was significant difference between SMg(i)(i = 1, ..., 33) and SMs(i)(i = 1, ..., 33), which meant that comparing to the self-snake based image enhancement method, higher similarity scores were obtained in the Gaussian filtering based image denoising. Their effect difference was 1.582, which might show that the image denoising would be more effective than the image enhancement methods for the performance improvement of the FRS.
In order to study the effects of the Gaussian filtering based image denoising on the recognition performance of the FRS under the conditions of different parameter settings. The repeated measurement analysis of variance method was carried out to compare the mean differences among the SM(i)(i = 1, ..., 33), SM1(i)(i = 1, ..., 33), SM2(i)(i = 1, ..., 33) and SM3(i)(i = 1, ..., 33). The descriptive statistics information were shown in Table 4.
Effect | Mean | Std. Deviation | N |
SM | 83.421 | 5.7417 | 33 |
SM 1 | 85.021 | 5.4986 | 33 |
SM 2 | 85.097 | 5.3577 | 33 |
SM 3 | 81.709 | 7.2233 | 33 |
The Greenhouse-Geisser method (P = 0.464) was used to analyze the tested data since they did not meet the Mauchly's Test of Sphericity (P = 0.000). The results showed that there were significant differences among SM(i)(i = 1, ..., 33), SM1(i)(i = 1, ..., 33), SM2(i)(i = 1, ..., 33) and SM3(i)(i = 1, ..., 33) (F(1.392, 44.536) = 15.096, P = 0.000). The subsequent pairwise comparison examinations were further carried out, and the results were shown in Table 5.
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SM1 | SM | 1.600* | 0.258 | 0.000 | 0.875 | 2.325 |
SM2 | −0.076 | 0.281 | 1.000 | −0.867 | 0.715 | |
SM3 | 3.312* | 0.816 | 0.002 | 1.016 | 5.608 | |
SM2 | SM | 1.676* | 0.347 | 0.000 | 0.701 | 2.650 |
SM1 | 0.076 | 0.281 | 1.000 | −0.715 | 0.867 | |
SM3 | 3.388* | 0.724 | 0.000 | 1.351 | 5.425 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
The results showed that there were significant differences between SM(i)(i = 1, ..., 33) vs. SM1(i)(i = 1, ..., 33) (P = 0.000) and SM2(i)(i = 1, ..., 33) (P = 0.000), which indicated that comparing to the original face images, the face images with the Gaussian filtering based image denoising with the Gaussian blur radius of the 0.5 and 1.0 pixels could both improve the recognition performance of the FRS, and their effects were 1.600 and 1.676 respectively. Nevertheless, the means between SM(i)(i = 1, ..., 33) and SM3(i)(i = 1, ..., 33) had no significant difference (P = 0.190).
Comparing to the Gaussian blur radius of the 2.0 pixels, the Gaussian filtering based image denoising with the parameter setting of the 0.5 and 1.0 pixels both gained the better recognition performance in the FRS, and their mean differences were 3.312 (P = 0.002) and 3.388 (P = 0.00) respectively. Nevertheless, the means between SM1(i)(i = 1, ..., 33) and SM2(i)(i = 1, ..., 33) had no significant difference (P = 1.000). The results indicated the Gaussian filtering based image denoising techniques could improve recognition performance, provided the correct parameters were applied. In our study, the quality of the questioned face images was dramatically worsened in the Gaussian filtering with the parameter setting of 2.0 pixels, as shown in Figure 4.
In order to study the effects of the different image denoising methods on the recognition performance of the FRS, we compared the effects of the image denoising methods among the Gaussian filtering (GF), the Wiener filtering (WF) [18], and the wavelet transform (WT) [19] based methods, which were all common used single-image denoising techniques in the FIHI. Considering the effects of different parameter settings on image denoising performance, we selected different combination of parameters for each used image denoising methods. The maximum scores of the face verification results under different parameters were served as the final scores of the corresponding denoising methods. In the Wiener filtering based image denoising, the parameters of the filtering window radius were set to 3.0 and 5.0 pixels, and the noise variances were set to 25 and 50 respectively, which finally resulted in four groups of denoising images in the Wiener filtering method. In the wavelet transform based method, the noise variances were specified as 0.2, 1.0, 2.0 and 4.0 respectively.
The repeated measurement analysis of variance method was applied to analyze the effects of the above image denoising methods on the recognition performance of the FSR. The means among the SMg(i)(i = 1, ..., 33), WF(i)(i = 1, ..., 33), WT(i)(i = 1, ..., 33) were compared and the descriptive statistics information were shown in Table 6.
Effects | Mean | Std. Deviation | Number |
SMg | 85.897 | 5.257 | 33 |
WF | 85.358 | 5.011 | 33 |
WT | 84.345 | 5.361 | 33 |
The pairwise comparison among SMg(i)(i = 1, ..., 33), WF(i)(i = 1, ..., 33), WT(i)(i = 1, ..., 33) showed that the significant differences were found between SMg(i)(i = 1, ..., 33) and WT(i)(i = 1, ..., 33) (P = 0.001), and mean difference was 1.552, which meant that comparing to the wavelet transform based image denoising method, the Gaussian filtering could significantly improve the recognition performance of the FSR. In addition, there was significant difference between WF(i)(i = 1, ..., 33) and WT(i)(i = 1, ..., 33) (P = 0.049), and mean difference was 1.012. No significant difference was observed between SMg(i)(i = 1, ..., 33) and WF(i)(i = 1, ..., 33) (P = 0.227).
In this paper, face images materials from 33 actual cases in forensic identification of human images were collected. The statistical effects of the image denoising and enhancement methods on face recognition were studied. The Gaussian filtering based method was used for image denoising. For the image enhancement, the self-snake model based technique was applied. The studied image processing methods are common and popular in the FIHI. The statistical data analysis techniques were used to quantitatively compute the effects of the different image processing methods on the recognition performance of the state-of-the-art neural network based face recognition system.
Our results found that both the Gaussian filtering based image denoising and the self-snake based image enhancement methods could improve the recognition performance of the used FRS. However, the effects of the Gaussian filtering based method were better than the later. This might be attributed to the facts that the image enhancement method improved the visual display quality of images, at the same time, it also partially augmented the image noises. In our study the questioned face images were all collected from the real cases. The image noises not only existed in the high frequency information of the images, but the face contour information might also partially be contaminated in the real cases. Furthermore, in order to be more robust to the variations of image quality, face angles, etc. Current neural network-based FRS have been paid more attention to the face overall outline information which mainly exists in the low frequency part of images. The face local detail information has relatively little effects on the recognition performance of the FRS. For example, in one of our studied cases, the suspect has a large spot on his right cheek. We removed this face spot by using image processing techniques. The similarity score in our used FRS between the original face and the face after removing the spot is still up to 99.999%. Our used Gaussian filtering based image denoising is a typical image smoothing method, which erases the contaminated high frequency information of images and then highlights the outline information of faces.
In addition, among our studied three image denoising methods, the effects of both the filtering based image denoising methods, i.e. the Gaussian filtering and the Wiener filtering based methods, were better than the wavelet transform based method which is a typical transform domain image denoising technique. Generally, spatial domain filters remove noises to a reasonable extent but at the cost of losing image sharp edges. However, the transform techniques in image denoising usually have the features that they remove noises while still preserving image characteristics, such as edges. The different effects of the spatial domain and transform domain image denoising techniques might be attributed to the fact that, in our study, the tested face images were all from real cases in the FIHI, which largely encountered the environmental corruption of noise, blur, and bad illumination. The high frequency components, such as edges and textures of face images, usually corrupted by noises. The Gaussian filtering and the Wiener filtering based methods removed noises in the spatial domain without specially focusing on image characteristic existing in the high frequency domain. The above features might be the key for the maximum effects of the spatial domain filters based denoising techniques on the performance improvement of the FRS.
To sum up, we found that face image quality could play a significant effect on the recognition performance of the face recognition system, and the image processing techniques could improve the recognition precision of the face recognition system. In addition, the effects of the Gaussian filtering were better than the self-snake model based image enhancement method, which indicated that the image denoising methods were more suitable for performance improvement of the deep face recognition system rather than the image enhancement techniques in the application of the practical cases. Our results have important guiding roles in the selection of image processing methods for deep face recognition applications in real cases of FIHI. The correct parameters for image processing are important in practical application. An automatic parameter selection framework will be valuable for performance improvement of the FRS. However, in the real cases of the FIHI, the traversing method of the parameters in a certain range will be very useful despite its simplicity, which will be studied in future work. Furthermore, we only studied the effects of the two representative image processing techniques, other state-of-the-art denoising and enhancement techniques, such as current hot methods of deep denoising, etc., would be further studied under real case applications in future work. In addition, the number of the tested face image materials from real cases of the FIHI was relatively few when compared with the public face databases. The face image set would be further extended in future work for more precise data analysis.
This work is supported by Shanghai Sailing Program (17YF1420000), and Ministry of Finance, PR China (GY2018G-6 and GY2020G-8).
The authors declare there is no conflict of interest.
[1] |
U. Scherhag, C. Rathgeb, J. Merkle, R. Breithaupt, C. Busch, Face recognition systems under morphing attacks: A survey, IEEE Access, 7 (2019), 23012–23026. doi: 10.1109/ACCESS.2019.2899367
![]() |
[2] |
A. Fydanaki, Z. Geradts, Evaluating OpenFace: an open-source automatic facial comparison algorithm for forensics, Forensic Sci. Res., 3 (2018), 202–209. doi: 10.1080/20961790.2018.1523703
![]() |
[3] | M. P. Evison, R. W. V. Bruegge, The magna database: A database of three-dimensional facial images for research in human identification and recognition, Forensic Sci. Commun., 10 (2008), 1–11. |
[4] | J. Zeng, S. Shi, Q. Lu, Y. Li, X. Qiu, Face recognition techniques for forensic identification of human images, Int. J. Digital Content Technol. Its Appl., 11 (2017), 34–43. |
[5] |
J. Han, B. Bhanu, Individual recognition using gait energy image, IEEE Trans. Pattern Anal. Mach. Intell., 28 (2006), 316–322. doi: 10.1109/TPAMI.2006.38
![]() |
[6] | S. Bharadwaj, H. Bhatt, M. Vatsa, R. Singh, A. Noore, Quality assessment based denoising to improve face recognition performance, CVPR 2011 WORKSHOPS, Colorado Springs, CO, 2011,140–145. |
[7] | A. A. Dawood, M. F. Saleh, Review of different techniques for image denoising, Int. J. Innovative Res. Comput. Commun. Eng., 6 (2018), 2498–2505. |
[8] |
L. Fan, F. Zhang, H. Fan, C. Zhang, Brief review of image denoising techniques, Visual Comput. Ind. Biomed. Art, 2 (2019), 7. doi: 10.1186/s42492-019-0016-7
![]() |
[9] |
P. Janani, J. Premaladha, K. S. Ravichandran, Image enhancement techniques: A study, Indian J. Sci. Technol., 8 (2015), 125–145. doi: 10.17485/ijst/2015/v8iS6/71216
![]() |
[10] | G. Deng, L. W. Cahill, An adaptive Gaussian filtering for noise reduction and edge detection, Nuclear Science Symposium and Medical Imaging Conference, 3 (1993), 1615–1619. |
[11] | X. Jia, S. Liu, X. Feng,, L. Zhang, FOCNet: A fractional optimal control network for image denoising, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Available from: https://ieeexplore.ieee.org/document/8954104. |
[12] | C. Tian, Y. Xu, L. Fei, K. Yan, Deep learning for image denoising: a survey, International Conference on Genetic and Evolutionary Computing, 384 (2019), 563–572. |
[13] |
C. Knaus, M. Zwicker, Progressive image denoising, IEEE Trans. Image Process., 23 (2014), 3114–3125. doi: 10.1109/TIP.2014.2326771
![]() |
[14] | S. Cha, T. Moon, Fully convolutional pixel adaptive image denoiser, 2019 IEEE/CVF International Conference on Computer Vision (ICCV), 2019, 4160–4169. |
[15] | J. Xu, L. Zhang, D. Zhang, A trilateral weighted sparse coding scheme for real-world image denoising, The European Conference on Computer Vision (ECCV), 2018, 20–36. |
[16] | R. Wang, X. Xiao, B. Guo, Q. Qin, R. Chen, An effective image denoising method for UAV images via improved generative adversarial networks, Sensors, 18 (2018), 1985. |
[17] |
M. Diwakar, M. Kumar, A review on CT image noise and its denoising, Biomed. Signal Process. Control, 42 (2018), 73–88. doi: 10.1016/j.bspc.2018.01.010
![]() |
[18] | J. Benesty, J. D. Chen, Y. T. Huang, Study of the widely linear wiener filter for noise reduction, IEEE International Conference on Acoustics, Speech, and Signal Processing, Dallas, Texas, USA, 2010,205–208. |
[19] |
S. G. Mallat, A theory for multiresolution signal decomposition: the wavelet representation, IEEE Trans. Pattern. Anal. Mach. Intell., 11 (1989), 674–693. doi: 10.1109/34.192463
![]() |
[20] | W. Zhang, R. Li, K. He, J. Zhou, Approach to image denoising and enhancement based on 3DDCT, J. Univ. Electron. Sci. Technol. China, 40 (2011), 742–746. |
[21] | T. Yan, M. Qu, C. Zhou, Image denoising with BEMD and edge-preserving self-snake model, International Conference on Intelligent Computing, 8588 (2014), 435–442. |
[22] |
K. Nguyen, C. Fookes, S. Sridharan, M. Tistarelli, M. Nixon, Super-resolution for biometrics: A comprehensive survey, Pattern Recognit., 78 (2018), 23–42. doi: 10.1016/j.patcog.2018.01.002
![]() |
[23] | J. Jiang, J. Ma, C. Chen, X. Jiang, Z. Wang, Noise robust face image super-resolution through smooth sparse representation, IEEE Trans. Cybernetics, 47 (2016), 3991–4002. |
[24] | R. Tkachenko, P. Tkachenko, I. Izonin, Y. Tsymbal, Learning-based image scaling using neural-like structure of geometric transformation paradigm, in Advances in Soft Computing and Machine Learning in Image Processing, (2018), 537–565. |
[25] | I. Izonin, R. Tkachenko, D. Peleshko, T. Rak, D. Batyuk, Learning-based image super-resolution using weight coefficients of synaptic connections, 2015 Xth International Scientific and Technical Conference "Computer Sciences and Information Technologies" (CSIT), 2015, 25–29. |
[26] | Y. Rashkevych, D. Peleshko, O. Vynokurova, I. Izonin, N. Lotoshynska, Single-frame image super-resolution based on singular square matrix operator, 2017 IEEE First Ukraine Conference on Electrical and Computer Engineering, 2017,944–948. |
[27] | M. Emambakhsh, J. Gao, A.N. Evans, An evaluation of denoising algorithms for 3D face recognition, 5th International Conference on Imaging for Crime Detection and Prevention, London, 2013. |
[28] | L. Zhang, L. Zhang, L. Li, Illumination quality assessment for face images: A benchmark and a convolutional neural networks based model, in International Conference on Neural Information Processing, Springer, Cham, 2017,583–593. |
[29] |
N. Zhuang, Q. Zhang, C. Pan, B. Ni, Y. Xu, X. Yang, W. Zhang, Recognition oriented facial image quality assessment via deep convolutional neural network, Neurocomputing, 358 (2019), 109–118. doi: 10.1016/j.neucom.2019.04.057
![]() |
[30] |
J. Yu, K. Sun, F. Gao, S. Zhu, Face biometric quality assessment via light CNN, Pattern Recog. Lett., 107 (2018), 25–32. doi: 10.1016/j.patrec.2017.07.015
![]() |
[31] | V. V. Starovoitov, D. I. Samal, D. V. Briliuk, Image enhancement for face recognition, International Conference on Iconics, St. Petersburg, Russia, 2003. |
[32] |
T. Ojala, M. Pietikainen, T. Maenpaa, Multiresolution gray-scale and rotation invariant texture classification with local binary patterns, IEEE Trans. Pattern Anal. Mach. Intell., 24 (2002), 971–987. doi: 10.1109/TPAMI.2002.1017623
![]() |
[33] |
S. Chang, B. Yu, M. Vetterli, Adaptive wavelet thresholding for image denoising and compression, IEEE Trans. Image Process., 9 (2000), 1532–1546. doi: 10.1109/83.862633
![]() |
[34] | D. Andrews, C. Mallows, Scale mixtures of normal distributions, J. Royal Stat. Soc., 36 (1974), 99–102. |
[35] | A. Martinez, R. Benevento, The AR face database, CVC Technical Report #24, 1998. |
[36] |
P. J. Phillips, P. J. Flynn, T. Scruggs, K. W. Bowyer, J. Chang, K. Hoffman, et al., Overview of the face recognition grand challenge, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), 1(2005), 947–954. doi: 10.1109/CVPR.2005.268
![]() |
[37] | S. Mittal, S. Agarwal, M. J. Nigam, Real time multiple face recognition: A deep learning approach, Proceedings of the 2018 International Conference on Digital Medicine and Image Processing, 2018, 70–76. |
[38] | I. Masi, Y. Wu, T. Hassner, P. Natarajan, Deep face recognition: A survey, 2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), 2018. Available from: https://ieeexplore.ieee.org/document/8614364. |
[39] |
G. Gilboa, N. Sochen, Y. Y. Zeevi, Image enhancement and denoising by complex diffusion process, IEEE Trans. Patt. Anal. Mach. Intell., 26 (2004), 1020–1036. doi: 10.1109/TPAMI.2004.47
![]() |
[40] |
P. Perona, J. Malik, Scale space and edge detection using anisotropic diffusion, IEEE Trans. Pattern Anal. Mach. Intell., 12 (1990), 629–639. doi: 10.1109/34.56205
![]() |
1. | Eissa Alreshidi, Rabie A. Ramadan, Md. Haidar Sharif, Omer Faruk Ince, Ibrahim Furkan Ince, A Comparative Study of Image Descriptors in Recognizing Human Faces Supported by Distributed Platforms, 2021, 10, 2079-9292, 915, 10.3390/electronics10080915 | |
2. | Rucha Golwalkar, Ninad Mehendale, Masked-face recognition using deep metric learning and FaceMaskNet-21, 2022, 52, 0924-669X, 13268, 10.1007/s10489-021-03150-3 | |
3. | Jinhua Zeng, Xiulian Qiu, Qimeng Lu, Huaping Zhu, Shaopei Shi, 2022, Chapter 89, 978-3-030-81006-1, 782, 10.1007/978-3-030-81007-8_89 | |
4. | Divyagna Bavikadi, A. Manjunatha, Abhishek Pol, Akshat Kadam, Prajakta Kulkarni, Aparna Singh, P. M. Kamble, Ravindra Hegadi, 2022, Chapter 13, 978-3-031-07004-4, 134, 10.1007/978-3-031-07005-1_13 | |
5. | Guoqiang Deng, Min Tang, Yuhao Zhang, Ying Huang, Xuefeng Duan, Privacy-Preserving Outsourced Artificial Neural Network Training for Secure Image Classification, 2022, 12, 2076-3417, 12873, 10.3390/app122412873 | |
6. | Wenting Liu, Yi Zuo, Tieshan Li, C.L. Philip Chen, 2022, Application of Image Recognition for On-board Sailor Behavior Based on Broad Learning System, 978-1-6654-5982-2, 1, 10.1109/DOCS55193.2022.9967770 | |
7. | K. Meena, J. N. Swaminathan, T. Rajendiran, S. Sureshkumar, N. Mohamed Imtiaz, 2022, Chapter 17, 978-981-16-7166-1, 225, 10.1007/978-981-16-7167-8_17 | |
8. | Shefali Arora, Ruchi Mittal, Harshita Kukreja, M. P. S. Bhatia, An evaluation of denoising techniques and classification of biometric images based on deep learning, 2023, 82, 1380-7501, 8287, 10.1007/s11042-021-11573-w | |
9. | Parama Bagchi, Debotosh Bhattacharjee, JULive3D: a live image acquisition protocol for real-time 3D face recognition, 2024, 83, 1380-7501, 1841, 10.1007/s11042-023-15728-9 | |
10. | Md. Abu Marjan, Mahmudul Hasan, Md. Zahidul Islam, Md. Palash Uddin, Masud Ibn Afjal, 2022, Masked Face Recognition System using Extended VGG-19, 979-8-3503-2054-1, 1, 10.1109/ICECTE57896.2022.10114484 | |
11. | Vrisha Sheth, Anya Priyal, Kavya Mehta, Nirali Desai, Manan Shah, Schematized study for tackling COVID-19 with Machine Learning (ML), Artificial Intelligence (AI), and Internet of Things (IoT), 2024, 2949866X, 10.1016/j.ipha.2024.04.003 | |
12. | Mengxuan Yang, Shengnan Li, Jinhua Zeng, The Effects of AI-Driven Face Restoration on Forensic Face Recognition, 2024, 14, 2076-3417, 3783, 10.3390/app14093783 | |
13. | Pei Jiang, Takashi Obi, Yoshikazu Nakajima, Integrating prior knowledge to build transformer models, 2024, 16, 2511-2104, 1279, 10.1007/s41870-023-01635-7 | |
14. | Yanwei Ren, Ridong Zhang, Furong Gao, A network structure for industrial process fault diagnosis based on hyper feature extraction and stacked LSTM, 2024, 287, 00092509, 119745, 10.1016/j.ces.2024.119745 | |
15. | S.V. Vasantha, B. Kiranmai, Mohammed Ashraf Hussain, Syeda Sufia Hashmi, Leema Nelson, S. Hariharan, 2023, Face and Object Detection Algorithms for People Counting Applications, 979-8-3503-4023-5, 1188, 10.1109/ICACRS58579.2023.10405114 | |
16. | Ali Ercetin, Oguzhan Der, Fatih Akkoyun, Manjunath Patel Gowdru Chandrashekarappa, Ramazan Şener, Mücahit Çalışan, Nevzat Olgun, Ganesh Chate, Kurki Nagaraja Bharath, Review of Image Processing Methods for Surface and Tool Condition Assessments in Machining, 2024, 8, 2504-4494, 244, 10.3390/jmmp8060244 | |
17. | Mohammad Fauzi Aziz, Gavin Sadiya Taraka, Sidharta Sidharta, Face Recognition as Base Protocol in Online Transactions, 2024, 245, 18770509, 166, 10.1016/j.procs.2024.10.240 | |
18. | Isma Batool, Muhammad Imran, A dual residual dense network for image denoising, 2025, 147, 09521976, 110275, 10.1016/j.engappai.2025.110275 | |
19. | Yue Zhao, Jianjun Bai, Hongbo Zou, Jing Feng, Multi‐scale feature fusion network‐based industrial process fault diagnosis method using space–time capsule and classifier optimization, 2025, 0008-4034, 10.1002/cjce.25682 | |
20. | Parul Choudhary, Pooja Pathak, Phalguni Gupta, Physiological biometric image quality assessment -a review, 2025, 1573-7721, 10.1007/s11042-025-20830-1 |
SizeNo | Mean | Std. Deviation | N |
Size1 | 80.142 | 4.8286 | 12 |
Size2 | 84.255 | 4.7798 | 11 |
Size3 | 86.440 | 6.1661 | 10 |
Total | 83.421 | 5.7417 | 33 |
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SM | SMg | −2.445 * | 0.309 | 0.000 | −3.226 | −1.665 |
SMs | −0.864 * | 0.193 | 0.000 | −1.351 | −0.377 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SMg | SM | −2.445 * | 0.309 | 0.000 | −3.226 | −1.665 |
SMs | 1.582 * | 0.304 | 0.000 | 0.813 | 2.350 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
Effect | Mean | Std. Deviation | N |
SM | 83.421 | 5.7417 | 33 |
SM 1 | 85.021 | 5.4986 | 33 |
SM 2 | 85.097 | 5.3577 | 33 |
SM 3 | 81.709 | 7.2233 | 33 |
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SM1 | SM | 1.600* | 0.258 | 0.000 | 0.875 | 2.325 |
SM2 | −0.076 | 0.281 | 1.000 | −0.867 | 0.715 | |
SM3 | 3.312* | 0.816 | 0.002 | 1.016 | 5.608 | |
SM2 | SM | 1.676* | 0.347 | 0.000 | 0.701 | 2.650 |
SM1 | 0.076 | 0.281 | 1.000 | −0.715 | 0.867 | |
SM3 | 3.388* | 0.724 | 0.000 | 1.351 | 5.425 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
Effects | Mean | Std. Deviation | Number |
SMg | 85.897 | 5.257 | 33 |
WF | 85.358 | 5.011 | 33 |
WT | 84.345 | 5.361 | 33 |
SizeNo | Mean | Std. Deviation | N |
Size1 | 80.142 | 4.8286 | 12 |
Size2 | 84.255 | 4.7798 | 11 |
Size3 | 86.440 | 6.1661 | 10 |
Total | 83.421 | 5.7417 | 33 |
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SM | SMg | −2.445 * | 0.309 | 0.000 | −3.226 | −1.665 |
SMs | −0.864 * | 0.193 | 0.000 | −1.351 | −0.377 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SMg | SM | −2.445 * | 0.309 | 0.000 | −3.226 | −1.665 |
SMs | 1.582 * | 0.304 | 0.000 | 0.813 | 2.350 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
Effect | Mean | Std. Deviation | N |
SM | 83.421 | 5.7417 | 33 |
SM 1 | 85.021 | 5.4986 | 33 |
SM 2 | 85.097 | 5.3577 | 33 |
SM 3 | 81.709 | 7.2233 | 33 |
A | B | Mean Difference (A-B) | Std. Error | Sig 1 | 95% Confidence Interval for Difference 1 | |
Lower Bound | Upper Bound | |||||
SM1 | SM | 1.600* | 0.258 | 0.000 | 0.875 | 2.325 |
SM2 | −0.076 | 0.281 | 1.000 | −0.867 | 0.715 | |
SM3 | 3.312* | 0.816 | 0.002 | 1.016 | 5.608 | |
SM2 | SM | 1.676* | 0.347 | 0.000 | 0.701 | 2.650 |
SM1 | 0.076 | 0.281 | 1.000 | −0.715 | 0.867 | |
SM3 | 3.388* | 0.724 | 0.000 | 1.351 | 5.425 | |
*Note: 1 Adjustment for multiple comparisons: Bonferroni. * The mean difference is significant at the 0.05 level. |
Effects | Mean | Std. Deviation | Number |
SMg | 85.897 | 5.257 | 33 |
WF | 85.358 | 5.011 | 33 |
WT | 84.345 | 5.361 | 33 |