Journal Menu
Archive
Last Edition
Journal information

Vol.1, No.3, 2022: pp.98-107

USE OF FACE RECOGNITION SOFTWARE BY KARHUNEN LOVE METHOD

Authors:

Hadžib Salkić 1

, Marija Kvasina2

, Almira Salkić3

, Vladica Ristić4

1University College “CEPS -“Center for Business Studies”, Kiseljak, Bosnia and Herzegovina
2University “VITEZ”, Faculty of informational technology, Travnik, Bosnia and Herzegovina
3Center for Education “Algebra znanja”, Travnik, Bosnia and Herzegovina
4Unviersity Metropolitan, Faculty of Applied Ecology “Futura”, Belgrade, Serbia

Received: 03.05.2022.
Accepted: 25.08.2022.
Available: 30.09.2022.

Abstract:

Numerical simulations and checks of face recognition software on given image databases represent a type of empirical research. Face recognition software works on the principle of comparing a photo of the person’s face with the photos in the database. The operation of face recognition software can be divided into three stages. The first stage is face detection, the second stage is face tracking and the third stage is face recognition. For this purpose, software solutions have been developed, with different work techniques. However, it is characteristic that regardless of the different techniques, each expresses its effect with a probability expressed in percentages. Simply put, for now, no software solution can be said to be 100% effective. For now, no computer solution can be compared to the human ability to recognize and identify a person.

Keywords:

Software, database, face recognition, face detection, Karhunen Loeve

References:

[1] R. Rahim, T. Afriliansyah, H. Winata, D. Nofriansyah, Ratnadewi, S. Aryza, Research of Face Recognition with Fisher Linear Discriminant. IOP Conference Series: Materials Science and Engineering, 300, 2018: 012037.
https://dx.doi.org/10.1088/1757-899X/300/1/012037
[2] L.S. Khoo, M.S. Mahmood, Application of Facial Recognition Technology on Identification of the Dead During Large Scale Disasters. Forensic Science International: Synergy, 2, 2020: 238-239. https://dx.doi.org/10.1016/j.fsisyn.2020.07.001
[3] M. Hernandez-de-Menendez, R. MoralesMenendez, C.A. Escobar, J. Arinez, Biometric applications in education. International Journal on Interactive Design and Manufacturing (IJIDeM), 15(2-3), 2021: 365-80.
https://dx.doi.org/10.1007/s12008-021-00760-6
[4] I. McAteer, A. Ibrahim, G. Zheng, W. Yang, C. Valli, Integration of Biometrics and Steganography: A Comprehensive Review. Technologies, 7(2), 2019: 34. https://doi.org/10.3390/technologies7020034
[5] P. Kasprowski, Z. Borowska, K. Harezlak, Biometric Identification Based on Keystroke Dynamics. Sensors, 22(9), 2022: 3158. https://doi.org/10.3390/s22093158
[6] P. Li, L. Prieto, D. Mery, P. Flynn, Face recognition in low quality images: A survey. arXiv, 20218: 1-15. https://doi.org/10.48550/arXiv.1805.11519
[7] Y. Spreeuwers, R.N. Veldhuis, Low-resolution face recognition and the importance of proper alignment. IET Biometrics, 8(4), 2019: 267-276. https://doi.org/10.1049/iet-bmt.2018.5008
[8] Ž. Dženopoljac, Expertise of the digital record. Association of Forensic Experts for Information Technologies – IТ veštak, Beograd, Srbija, 2012: 1-9.
[9] M. Sajjad, M.Nasir, K. Muhammad, S. Khan, Z. Jan, A.K. Sangaiah, M. Elhoseny, S. W. Baik, Raspberry Pi assisted face recognition framework for enhanced law-enforcement services in smart cities. Future Generation Computer Systems, 108, 2020: 995-1007. https://doi.org/10.1016/j.future.2017.11.013
[10] C. Mingtsung, Q. Wei, H. Jiaqi, Z. Zhuomin, Research on the application of face recognition system. Journal of Physics: Conference Series, 1684, 2020: 012126. https://dx.doi.org/10.1088/1742-6596/1684/1/012126
[11] P. Kaur, K. Krishan, S.K. Sharma, T. Kanchan, Facial-recognition algorithms: A literature review. Medicine, Science and the Law, 60(2), 2020: 131-139. https://dx.doi.org/10.1177/0025802419893168
[12] N. van Rijsbergen, K. Jaworska, G.A. Rousselet, P.G. Schyns With age comes representational wisdom in social signals. Current Biology, 24(23), 2014: 2792-2796. https://dx.doi.org/10.1016/j.cub.2014.09.075
[13] R.E. Jack, P.G. Schyns, The Human Face as a Dynamic Tool for Social Communication. Current Biology, 25(14), 2015: R621-R634. https://doi.org/10.1016/j.cub.2015.05.052
[14] L.F. Barrett, R. Adolphs, S. Marsella, A.M. Martinez, S.D. Pollak Emotional Expressions Reconsidered: Challenges to Inferring Emotion from Human Facial Movements. Psychological Science in the Public Interest, 20(1), 2019: 1-68. https://dx.doi.org/10.1177/1529100619832930
[15] S. Gu, F. Wang, N.P. Patel, J.A. Bourgeois, J.H. Huang, A Model for Basic Emotions Using Observations of Behavior in Drosophila. Frontiers in Psychology, 10, 2019: 781. https://dx.doi.org/10.3389/fpsyg.2019.00781
[16] J. Zhao, M. Zhang, C. He K. Zuo, Data-Driven Research on the Matching Degree of Eyes, Eyebrows and Face Shapes. Frontiers in Psychology, 10, 2019:1 466. https://dx.doi.org/10.3389/fpsyg.2019.01466
[17] R. Lionnie, C. Apriono, D. Gunawan, Eyes versus Eyebrows: A Comprehensive Evaluation Using the Multiscale Analysis and Curvature-Based Combination Methods in Partial Face Recognition. Algorithms, 15(6), 2022: 208. https://doi.org/10.3390/a15060208
[18] S.Z. Li, A.K. Jain, Handbook of face recognition. Springer, London, 2011. https://doi.org/10.1007/978-0-85729-932-1
[19] K.H. Teoh, R.C. Ismail, S.Z.M. Naziri, R. Hussin, M.N.M. Isa, M.S.S.M. Basir, Face Recognition and Identification using Deep Learning Approach. Journal of Physics: Conference Series, 1755, 2021: 012006.
https://dx.doi.org/10.1088/1742-6596/1755/1/012006
[20] K. Karhunen, On Linear Methods in Probability Theory. The RAND Corp. Santa Monica, California, USA, 1960, T-131.
[21] T.H. Le, Applying Artificial Neural Networks for Face Recognition. Hindawi Publishing Corporation, Ho Chi Minh City, Vietnam, 2011.
[22] S. Alambdy, L. Elrefaiei, Deep Convolutional Neural Network-Based Approaches for Face Recognition. Applied Sciences, 9(20), 2019: 4397. https://doi.org/10.3390/app9204397
[23] Y.X. Yang, C. Wen, K. Xie, F.Q. Wen, G.Q. Sheng, X.G. Tang, Face Recognition Using the SR-CNN Model. Sensors, 18(12), 2018: 4237. https://doi.org/10.3390/s18124237
[24] S. Gudivada, B. Adrian, Face recognition using ortho-diffusion bases. 20th European Signal Processing Conference (EUSIPCO 2012), August 27-31, 2012, Bucharest, Romania, pp.1578-1782.
[25] D. Alonso, Science-driven 3D data compression. Monthly Notices of the Royal Astronomical Society, 473(4), 2018: 4306- 4317. https://doi.org/10.1093/mnras/stx2644
[26] S. Kumar, V. Dhadwal, A block Karhunen-Loève Transform based efficient lossy compression algorithm for optical remote sensing imagery. Geocarto International, 32(10), 2016: 1159-1170. https://doi.org/10.1080/10106049.2016.1195887
[27] R.D. Dony, K.R. Rao, P.C. Yip, Karhunen-Loève Transform: The Transform and Data Compression Handbook. CRC Press LLC, 2001.

© 2022 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0)

Volume 3
Number 1
March 2024.

 

Loading

How to Cite

H. Salkić, M. Kvasina, A. Salkić, V. Ristić, Use of Face Recognition Software by Karhunen Love Method. Advanced Engineering Letters, 1(3), 2022: 98–107.
https://doi.org/10.46793/adeletters.2022.1.3.4

More Citation Formats

Salkić, H., Kvasina, M., Salkić, A., & Ristić, V. (2022). Use of Face Recognition Software by Karhunen Love Method. Advanced Engineering Letters1(3), 98–107. https://doi.org/10.46793/adeletters.2022.1.3.4

Salkić, Hadžib, et al. “Use of Face Recognition Software by Karhunen Love Method.” Advanced Engineering Letters, vol. 1, no. 3, 2022, pp. 98–107, https://doi.org/10.46793/adeletters.2022.1.3.4.

Salkić, Hadžib, Marija Kvasina, Almira Salkić, and Vladica Ristić. 2022. “Use of Face Recognition Software by Karhunen Love Method.” Advanced Engineering Letters 1 (3): 98–107. https://doi.org/10.46793/adeletters.2022.1.3.4.

Salkić, H., Kvasina, M., Salkić, A. and Ristić, V. (2022). Use of Face Recognition Software by Karhunen Love Method. Advanced Engineering Letters, 1(3), pp.98–107. doi: 10.46793/adeletters.2022.1.3.4.