Face Recognition Approach Via Deep and Machine Lea
Face Recognition Approach Via Deep and Machine Lea
(3)
1 Introduction
152
Wasit Journal for Pure Sciences Vol. (2) No. (3)
technology and ensure safety in different environments [2]. Face recognition, utilizing
traditional techniques, has long been a fundamental approach in the field of computer
vision and biometric identification. Before the advent of deep learning and advanced
algorithms, traditional face recognition methods relied on the extraction of handcrafted
features and the application of statistical classifiers. These techniques involved analyz-
ing facial characteristics such as shape, texture, and spatial relationships to establish
identity. Although traditional face recognition methods may not possess the same level
of accuracy and robustness as deep learning approaches, they have laid the foundation
for the development of modern face recognition systems [3]. With deep learning, face
recognition has revolutionized the field of computer vision and biometric identification.
With its ability to automatically extract and analyze intricate facial features, deep learn-
ing has meaningfully improved the accuracy and efficiency of face recognition systems.
Through the utilization of deep neural networks, these systems these systems have the
capacity to acquire knowledge and understand complex patterns, allowing for robust
identification and verification of individuals [4]. The deep wavelet scattering transform,
known for its ability to capture multi-scale and invariant representations, has emerged
as a promising approach for extracting robust features from facial images. By decom-
posing facial data into different frequency bands and orientations, the deep wavelet
scattering transform effectively captures both local and global information, enabling
enhanced face recognition accuracy [5]. The upcoming sections of this paper follow
this organization: Section 2 examines related literature, Section 3 outlines the research
background, Section 4 presents the methodology and the proposed approach in detail,
Section 5 analyzes the experimental results, and finally, Section 6 concludes the paper
along with discussing future work.
2 Related Literature
153
Wasit Journal for Pure Sciences Vol. (2) No. (3)
hances the network's visual clarity. The suggested model demonstrated exceptional per-
formance when compared to other CNN-based networks, achieving an impressive train-
ing accuracy of 85.76% and a validation accuracy of 64.40%. In 2020 Fahima Tabas-
sum, et al. [9] The combination of the Discrete Wavelet Transform (DWT) coherence
with four different algorithms, namely principal component analysis (PCA) error vec-
tor, PCA eigen vector, Linear Discriminant Analysis (LDA) eigen vector, and Convo-
lutional Neural Network (CNN), was utilized. The results of these four approaches were
fused using detection probability entropy and a Fuzzy system, resulting in an impres-
sive recognition rate of 89.56% even in the worst-case scenario. and an impressive
93.34% for the best-case scenario. In 2022 Showkat Ahmad Dar, et al. [10] developed
a novel Real Time Face Recognition (RTFR) system, comprising three primary stages.
In the initial phase, the system captures real-time video footage and collects 1056 face
images from 24 individuals using a camera with a resolution of 112*92. The subsequent
stage employs an efficient RTFR algorithm to recognize faces by comparing them
against a pre-existing database. This algorithm leverages two distinct deep learning ap-
proaches, namely CNN and VGG-16 with Transfer Learning, to improve the perfor-
mance of the RTFR system. The proposed algorithm demonstrates impressive accuracy,
yielding a result of 99.37.
3 Research Background
In this section, a brief explanation of the various tools (algorithms and techniques)
employed in this proposal will be provided.
154
Wasit Journal for Pure Sciences Vol. (2) No. (3)
face regions obtained from R-Net. It also performs additional face classification to de-
termine if the detected regions indeed contain faces.
Fig. 2. Operations of Wavelet Scatter where 𝐼 represents the input data, 𝜓 denotes a wavelet
function and 𝜑 represents an averaging low-pass filter.
Suppose we have an image (𝐼). The first scattering coefficient is the average of the
image and can be obtained by convolving the image with scaling filter (low pass filter)
𝜑𝑗 , as Equation 1 [14]:
𝑆0,𝑗 (𝐼) = 𝐼 ∗ 𝜑𝐽 (1)
Where 𝑗 a certain scale.
First-layer scattering coefficients are obtained by applying convolution to the input im-
age (𝐼) using a wavelet filter (𝜓𝜆1 ) at a certain scale (𝑗) and taking the modulus of the
155
Wasit Journal for Pure Sciences Vol. (2) No. (3)
resulting coefficients, followed by low-pass filtering with a scaling filter (𝜑𝑗 ). The co-
efficient is determined by Equation 2 [14]:
𝑆1,𝑗 (𝜆1 , 𝐼) = |𝐼 ∗ 𝜓𝜆1 | ∗ 𝜑𝑗 (𝐼) (2)
Second-layer scattering coefficients are obtained by convolving the image (I) by a
wavelet filter (𝜓𝜆1 ) at a certain scale (𝑗), taking the modulus of the resulting coeffi-
cients, followed by convolving again with another wavelet filter (𝜓𝜆2 ) at a different
scale and taking the modulus, and finally low-pass filtering with a scaling filter (𝜑𝑗 ).
The coefficient is determined by Equation 3 [14]:
𝑆2,𝑗 ((𝜆1 , 𝜆2 ), 𝐼) = ||𝐼 ∗ 𝜓𝜆1 | ∗ 𝜓𝜆2 | ∗ 𝜑𝑗 (𝐼) (3)
Note that the length of local translation invariance is determined by the parameter (𝑗),
which is the breadth of the low-pass filter. In addition to the amount of scales the trans-
form can produce, 𝜑 signifies a low-pass filter, 𝜓 denoted wavelet, 𝜆 is the rotation op-
erations, since 𝑆1 , 𝑗 and 𝑆2 , 𝑗 are the outputs of low-pass filters, they can be down-sam-
pled according to the filter width 2 power 𝑗. The last layer of the Wavelet Scattering
network is the average pooling layer, which computes the average of each scattering
coefficient over the spatial domain. The network's final output is a vector of all the
calculated averages, which captures the low-level and high-level texture features of the
input signal or image. Equation 4 determines the final layer's output [14]
156
Wasit Journal for Pure Sciences Vol. (2) No. (3)
SVM is a widely used supervised machine learning algorithm utilized for classifica-
tion and regression purposes. It is particularly renowned for binary classification tasks,
though it can be adapted for multi-class scenarios too. SVM is especially valued for its
proficiency in dealing with high-dimensional data and its capability to handle complex
datasets with well-defined margin separation. It operates by locating the optimal hyper-
plane in a high-dimensional space that divides the data into various classes [18].
4 Methodology
Through the use of a wavelet scattering transform network, we are investigating the
performance of face recognition in this study. This section explains in detail the pro-
posed model, which consists of four stages: 1) Image Acquisition,2) Pre-processing
stage, 3) Features Extraction stage, 4) Classification stage. Figure 4 shows the steps of
a proposed model.
157
Wasit Journal for Pure Sciences Vol. (2) No. (3)
Load Dataset
Image Acquisition
Splitting Dataset
Nonlinearity
Nonlinearity
Convolution
Convolution
Wavelet
Wavelet
Averaging
Averaging
FC
Feature vector
Classification
SoftMax Classifier
&
SVM Classifier
158
Wasit Journal for Pure Sciences Vol. (2) No. (3)
Within this specific section, the proposed model’s efficiency is evaluated using face
sub-dataset in MULB dataset showed in section 3.
3520 frontal-face images with a bit depth of 24 are included in the MLUB and are
divided among 176 participants. Images in (.jpg) format type were taken under various
bright lighting situations using a range of expressions, positions, and accessories. We
159
Wasit Journal for Pure Sciences Vol. (2) No. (3)
divided it into 70% for training (2464 images), 10% for validation (352 images), and
20% for testing (704).
Precision: It is an evaluation metric that gauges the proportion of true positive pre-
dictions (accurately predicted positive samples) out of all positive predictions made by
the model. It serves as an indicator of the model's capability to minimize false positives,
indicating how well it identifies positive instances correctly. It can be determined as
Equation 7 [7]:
𝑇𝑝𝑜𝑠
𝑃𝑟𝑒 = (7)
(𝑇𝑝𝑜𝑠 +𝐹𝑝𝑜𝑠 )
Recall: It quantifies the fraction of true positive predictions among all actual positive
samples present in the dataset. It evaluates the ability of model to avoid false negatives.
The recall can be determined as Equation 8 [7]:
𝑇𝑝𝑜𝑠
𝑅𝑒𝑐 = (8)
(𝑇𝑝𝑜𝑠 +𝐹𝑛𝑒𝑔 )
160
Wasit Journal for Pure Sciences Vol. (2) No. (3)
5.3 The Evaluation of the Proposed Model Using Various Evaluation Metrics
First, we evaluated the performance when extracting the image features from deep
wavelet scattering transform network followed by SoftMax as a classifier. Second, Im-
age features were extracted from the same network, and subsequently, SVM was em-
ployed as the classifier. After a number of training rounds, the best accuracy rate for
the testing model was found to be 97.87%, and the loss function was 0.0948 with Soft-
Max classifier. During classification with the SVM classifier, the highest accuracy rate
for the testing model was achieved, reaching 98.29%. Figure 6 and Figure 7 show the
accuracy and loss function respectively for proposed model with SoftMax classifier.
Table 2 show the evaluation metrics for proposed model. The chart in figure 8 show the
accuracy of the model with both classifiers. From Table 2 and Figure 8 we notice the
SVM classifier give better result for face recognition model.
161
Wasit Journal for Pure Sciences Vol. (2) No. (3)
Fig. 6. The Proposed Model Accuracy Fig. 7. The Proposed Model Loss func-
tion
98.4
98.2
Accuracy
98
97.8
97.6
Type of classifier
Fig. 8. The Relationship Between the Classifier and the Accuracy of the Model
In this paper, deep wavelet scattering transform network were applied for feature
extraction from face image and machine learning technique (SoftMax & SVM) for clas-
sification. According to the experimental results, the SVM classifier has demonstrated
superior performance compared to the SoftMax classifier. The proposed face recogni-
tion method has been thoroughly evaluated on the MULB face database, showing its
effectiveness. Specifically, the SVM classifier achieved an impressive recognition ac-
curacy of 98.29%, while the SoftMax classifier achieved a slightly lower accuracy of
97.87%. In our future research, we plan to incorporate face recognition into multimodal
biometric techniques, along with other types of biometric data that have demonstrated
distinct and discriminative features in face images.
162
Wasit Journal for Pure Sciences Vol. (2) No. (3)
7 Acknowledgment
8 References
[1] M. Hasan Abdulameer and R. Adnan Kareem, "Face Identification Approach Using Legendre
Moment and Singular Value Decomposition," International Journal Of Computing Digital
System, vol. 13, no. 1, pp. 389-398, 2021.
[2] Y. Kortli, M. Jridi, A. Al Falou, and M. Atri, "Face recognition systems: A survey," Sensors, vol.
20, no. 2, pp. 3-36, 2020.
[3] Z. Xie, J. Li, and H. Shi, "A Face Recognition Method Based on CNN," in Journal of Physics:
Conference Series, vol. 1395, no. 1, pp. 1-8: IOP Publisher, 2019.
[4] P. S. Prasad, R. Pathak, V. K. Gunjan, and H. Ramana Rao, "Deep learning based representation
for face recognition," in Proceedings of the 2nd International Conference on Communications and
Cyber Physical Engineering, pp. 419-424: Springer, 2020.
[5] A. Rehman, M. Harouni, M. Omidiravesh, S. M. Fati, and S. A. Bahaj, "Finger Vein
Authentication Based on Wavelet Scattering Networks," Computers, Materials Continua, vol. 72,
no. 2, pp. 3369-3383, 2022.
[6] N. H. Alskeini, K. N. Thanh, V. Chandran, and W. Boles, "Face recognition: Sparse representation
vs. Deep learning," in Proceedings of the 2nd International Conference on Graphics and Signal
Processing, pp. 31-37, 2018.
[7] S. Almabdy and L. Elrefaei, "Deep convolutional neural network-based approaches for face
recognition," Applied Sciences, vol. 9, no. 20, pp. 1-21, 2019.
[8] A. Gupta, S. Arunachalam, and R. Balakrishnan, "Deep self-attention network for facial emotion
recognition," Procedia Computer Science, vol. 171, pp. 1527-1534, 2020.
[9] F. Tabassum, M. I. Islam, R. T. Khan, and M. R. Amin, "Human face recognition with
combination of DWT and machine learning," Journal of King Saud University-Computer
Information Sciences, vol. 34, no. 3, pp. 546-556, 2022.
[10] S. A. Dar and S. Palanivel, "Performance Evaluation of Convolutional Neural Networks (CNNs)
And VGG on Real Time Face Recognition System," Advances in Science, Technology and
Engineering Systems Journal, vol. 6, no. 2, pp. 956-964, 2021.
[11] K. Zhang, Z. Zhang, Z. Li, and Y. Qiao, "Joint face detection and alignment using multitask
cascaded convolutional networks," IEEE signal processing letters, vol. 23, no. 10, pp. 1499-1503,
2016.
[12] S. Mallat, "Group invariant scattering," Communications on Pure Applied Mathematics, vol. 65,
no. 10, pp. 1331-1398, 2012.
[13] B. Soro and C. Lee, "A wavelet scattering feature extraction approach for deep neural network
based indoor fingerprinting localization," Sensors, vol. 19, no. 8, pp. 1-12, 2019.
[14] J. Andén and S. Mallat, "Deep scattering spectrum," IEEE Transactions on Signal Processing, vol.
62, no. 16, pp. 4114-4128, 2014.
[15] Z. Liu, G. Yao, Q. Zhang, J. Zhang, and X. Zeng, "Wavelet scattering transform for ECG beat
classification," Computational mathematical methods in medicine, pp. 1-11, 2020.
163
Wasit Journal for Pure Sciences Vol. (2) No. (3)
[16] P. Pandey, R. Singh, and M. Vatsa, "Face recognition using scattering wavelet under Illicit Drug
Abuse variations," in International Conference on Biometrics (ICB), pp. 1-6: IEEE, 2016.
[17] I. Kouretas and V. Paliouras, "Simplified hardware implementation of the softmax activation
function," in 8th international conference on modern circuits and systems technologies
(MOCAST), pp. 1-4: IEEE, 2019.
[18] C.-C. Chang and C.-J. Lin, "LIBSVM: A Library for Support Vector Machines," ACM
Transactions on Intelligent Systems and Technology, vol. 2, no. 3, pp. 1-27, 2011.
164