Abstract
Facial analysis evaluates the physical appearances of person, which is crucial for several clinical settings. The perspective of real-world faces captured in an uncontrolled environment makes it harder for the gender prediction algorithm to correctly identify gender. The accuracy of the most advanced algorithms currently in use for real-time facial gender prediction is decreased by these factors. Most importantly the facial gender prediction can pave a way for the visually challenged persons to identify the gender and age. A dual shot face detector with task restricted Fine-tuned deep neural network (DTFN) is created to recognise the facial land markings for accurate gender and age prediction in order to overcome the challenges and defects. Bidirectional filtering and sigmoid stretching are the main preprocessing methods used to improve contrast and remove noise from the input image once facial photographs are first gathered. Next, employing the modified dual shot face detector (DSFD) to separate the face from the remaining background image. To solve this problem, DSFD is built around caps net. A task constrained deep convolutional neural network (TCDCN) is then used to extract and identify features from facial landmarks. The collected features are fed into a fine tuned deep neural network (DNN) classifier, which further classifies the data according to age and gender. By adjusting the hidden layer’s parameter using the stochastic gradient descent technique, fine tweaking is achieved. According to the results of experimental research the proposed technique achieves 96\(\%\) accuracy, precision has 96.57\(\%\), and hit rate value of 96\(\%\). Thus, the proposed approach is the best option for automatic facial land mark detection.
REFERENCES
H. Jin, Sh. Liao, and L. Shao, ‘‘Pixel-in-pixel net: Towards efficient facial landmark detection in the wild,’’ Int. J. Comput. Vision 129, 3174–3194 (2021). https://doi.org/10.1007/s11263-021-01521-4
X. Guo, S. Li, J. Yu, J. Zhang, J. Ma, L. Ma, and H. Ling, ‘‘PFLD: A practical facial landmark detector,’’ arXiv Preprint (2019). https://doi.org/10.48550/arXiv.1902.10859
Zh. Liu, X. Zhu, G. Hu, H. Guo, M. Tang, Zh. Lei, N. M. Robertson, and J. Wang, ‘‘Semantic alignment: Finding semantically consistent ground-truth for facial landmark detection,’’ in 2019 IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), Long Beach, Calif., 2019 (IEEE, 2019), pp. 3467–3476. https://doi.org/10.1109/cvpr.2019.00358
M. Zhu, D. Shi, M. Zheng, and M. Sadiq, ‘‘Robust facial landmark detection via occlusion-adaptive deep networks,’’ in 2019 IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), Long Beach, 2019 (IEEE, 2019), pp. 3486–3496. https://doi.org/10.1109/cvpr.2019.00360
X. Zou, Sh. Zhong, L. Yan, X. Zhao, J. Zhou, and Yi. Wu, ‘‘Learning robust facial landmark detection via hierarchical structured ensemble,’’ in 2019 IEEE/CVF Int. Conf. on Computer Vision (ICCV), Seoul, 2019 (IEEE, 2019), pp. 141–150. https://doi.org/10.1109/iccv.2019.00023
J. Wan, Zh. Lai, J. Li, J. Zhou, and C. Gao, ‘‘Robust facial landmark detection by multiorder multiconstraint deep networks,’’ IEEE Trans. Neural Networks Learn. Syst. 33, 2181–2194 (2021). https://doi.org/10.1109/tnnls.2020.3044078
A. Asgarian, S. Zhao, A. B. Ashraf, M. E. Browne, K. M. Prkachin, A. Mihailidis, and B. Taati, ‘‘Limitations and biases in facial landmark detection—An empirical study on older adults with dementia,’’ in CVPR Workshop (2019), pp. 28–36.
Sh. Mane and G. Shah, ‘‘Facial recognition, expression recognition, and gender identification,’’ in Data Management, Analytics and Innovation, Ed. by V. Balas, N. Sharma, and A. Chakrabarti, Advances in Intelligent Systems and Computing, Vol. 808 (Springer, Singapore, 2019), pp. 275–290. https://doi.org/10.1007/978-981-13-1402-5_21
Z. Ferková, P. Urbanová, D. Černý, M. Žuži, and P. Matula, ‘‘Age and gender-based human face reconstruction from single frontal image,’’ Multimedia Tools Appl. 79, 3217–3242 (2020). https://doi.org/10.1007/s11042-018-6869-5
X. Dong and Yi. Yang, ‘‘Teacher supervises students how to learn from partially labeled images for facial landmark detection,’’ in 2019 IEEE/CVF Int. Conf. on Computer Vision (ICCV), Seoul, 2019 (IEEE, 2019), pp. 783–792. https://doi.org/10.1109/iccv.2019.00087
S. Sivachandiran, K. J. Mohan, and G. M. Nazer, ‘‘Automated deep learning based age and gender classification model using facial features for video surveillance,’’ J. Algebraic Stat. 13, 621–633 (2022). https://doi.org/10.52783/jas.v13i2.205
S. Haseena, S. Saroja, R. Madavan, A. Karthick, B. Pant, and M. Kifetew, ‘‘Prediction of the age and gender based on human face images based on deep learning algorithm,’’ Comput. Math. Methods Med. 2022, 1413597 (2022). https://doi.org/10.1155/2022/1413597
M. D. A. Angeloni and H. Pedrini, ‘‘Improving makeup face verification by exploring part-based representations,’’ arXiv Preprint (2021). https://doi.org/10.48550/arXiv.2101.07338
S. T. Rahman, A. Arefeen, Sh. Sh. Mridul, A. I. Khan, and S. Subrina, ‘‘Human age and gender estimation using facial image processing,’’ in 2020 IEEE Region 10 Symp. (TENSYMP), Dhaka, Bangladesh, 2020 (IEEE, 2020), pp. 1001–1005. https://doi.org/10.1109/tensymp50017.2020.9230933
V. Raman, Kh. Elkarazle, and P. Then, ‘‘Gender-specific facial age group classification using deep learning,’’ Intell. Autom. Soft Comput. 34, 105–118 (2022). https://doi.org/10.32604/iasc.2022.025608
H. Kim, H. Kim, and E. Hwang, ‘‘Real-time facial feature extraction scheme using cascaded networks,’’ in 2019 IEEE Int. Conf. on Big Data and Smart Computing (BigComp), Kyoto, Japan, 2019 (IEEE, 2019), pp. 1–7. https://doi.org/10.1109/bigcomp.2019.8679316
A. B. A. Hassanat, A. A. Albustanji, A. S. Tarawneh, M. Alrashidi, H. Alharbi, M. Alanazi, M. Alghamdi, I. S. Alkhazi, and V. B. S. Prasath, ‘‘DeepVeil: Deep learning for identification of face, gender, expression recognition under veiled conditions,’’ Int. J. Biometrics 14, 453–480 (2022). https://doi.org/10.1504/ijbm.2022.124683
A. Gurnani, K. Shah, V. Gajjar, V. Mavani, and Ya. Khandhediya, ‘‘SAF-BAGE: Salient approach for facial soft-biometric classification—Age, gender, and facial expression,’’ in 2019 IEEE Winter Conf. on Applications of Computer Vision (WACV), Waikoloa, Hawaii, 2019 (IEEE, 2019), pp. 839–847. https://doi.org/10.1109/wacv.2019.00094
H.-W. Kim, H.-J. Kim, S. Rho, and E. Hwang, ‘‘Augmented EMTCNN: A fast and accurate facial landmark detection network,’’ Appl. Sci. 10, 2253 (2020). https://doi.org/10.3390/app10072253
P. Chandran, D. Bradley, M. Gross, and T. Beeler, ‘‘Attention-driven cropping for very high resolution facial landmark detection,’’ in 2020 IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), Seattle, Wash., 2020 (IEEE, 2020), pp. 5861–5870. https://doi.org/10.1109/cvpr42600.2020.00590
J. Chen, Sh. Xu, J. Zou, and Z. Chen, ‘‘Interrupted-sampling repeater jamming suppression based on stacked bidirectional gated recurrent unit network and infinite training,’’ IEEE Access 7, 107428–107437 (2019). https://doi.org/10.1109/access.2019.2932793
M. Waheed, S. Allaoua Chelloug, M. Shorfuzzaman, A. Alsufyani, A. Jalal, Kh. Alnowaiser, and J. Park, ‘‘Exploiting human pose and scene information for interaction detection,’’ Comput., Mater. Continua 74, 5853–5870 (2023). https://doi.org/10.32604/cmc.2023.033769
D. I. J. Jacob, ‘‘Performance evaluation of caps-net based multitask learning architecture for text classification,’’ J. Artif. Intell. Capsule Networks 2 (1), 1–10 (2020). https://doi.org/10.36548/jaicn.2020.1.001
T. Zhu, Ch. Zhang, T. Wu, Zh. Ouyang, H. Li, X. Na, J. Liang, and W. Li, ‘‘Research on a real-time driver fatigue detection algorithm based on facial video sequences,’’ Appl. Sci. 12, 2224 (2022). https://doi.org/10.3390/app12042224
Yu. Lei and Yi. Ying, ‘‘Fine-grained analysis of stability and generalization for stochastic gradient descent,’’ Proc. Mach. Learn. Res. 119, 5809–5819 (2020). https://proceedings.mlr.press/v119/lei20c.html.
Funding
This work was supported by ongoing institutional funding. No additional grants to carry out or direct this particular research were obtained.
Author information
Authors and Affiliations
Contributions
The corresponding author claims the major contribution of the paper including formulation, analysis and editing. The co-author provides guidance to verify the analysis result and manuscript editing.
Corresponding authors
Ethics declarations
The authors declared that they have no conflicts of interest to this work. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.
Additional information
Publisher’s Note.
Allerton Press remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Meenakshi, J., Thailambal, G. Modified DSFD and TCDCN Based Facial Landmark Detection for Gender and Age Classification. Optoelectron.Instrument.Proc. 60, 398–411 (2024). https://doi.org/10.3103/S8756699024700468
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.3103/S8756699024700468