Abstract
Automated human posture estimation (A-HPE) systems need delicate methods for detecting body parts and selecting cues based on marker-less sensors to effectively recognize complex activity motions. Recognition of human activities using vision sensors is a challenging issue due to variations in illumination conditions and complex movements during the monitoring of sports and fitness exercises. In this paper, we propose a novel A-HPE method that intelligently identifies human behaviours by utilizing saliency silhouette detection, robust body parts model and multidimensional cues from full-body silhouettes followed by an entropy Markov model. Initially, images are pre-processed and noise is removed to obtain a robust silhouette. Body parts models are then used to extract twelve key body parts. These key body parts are further optimized to assist the generation of multidimensional cues. These cues include energy, optical flow and distinctive values that are fed into quadratic discriminant analysis to discriminate cues which help in the recognition of actions. Finally, these optimized patterns are further processed by a maximum entropy Markov model as a recognizer engine based on transition and emission probability values for activity recognition. For evaluation, we used a leave-one-out cross validation scheme and the results outperformed existing well-known statistical state-of-the-art methods by achieving better body parts detection and higher recognition accuracy over four benchmark datasets. The proposed method will be useful for man-machine interactions such as 3D interactive games, virtual reality, service robots, e-health fitness, and security surveillance.
Design model of automatic posture estimation and action recognition.
Similar content being viewed by others
References
Ahmed A, Jalal A, Kim K (2020) A novel statistical method for scene classification based on multi–object categorization and logistic regression. Sensors
Al-Ghannam R, Al–Dossari H (2016) Prayer Activity Monitoring and Recognition Using Acceleration Features with Mobile Phone. Arabian J Sci Eng 41:4967–4979
Andriluka M, Pishchulin L, Gehler P, Schiele (2014) 2D Human Pose Estimation: New Benchmark and State of the Art Analysis. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/CVPR.2014.471
Badar S, Jalal A, Kim K (2020) Wearable Inertial Sensors for Daily Activity Analysis Based on Adam Optimization and the Maximum Entropy Markov Model. Entropy 22(5):1–19
Bay H, Tuytelaars T, Gool LV (2006) SURF: Speeded up robust features.European Conference of Computer Vision. https://doi.org/10.1007/11744023
Beigi H (2010) Voice: technologies and algorithms for biometrics applications Homayoon Beigi. IEEE Courses: Bioengineering
Chen H, Mcgurr M (2014) Improved color and intensity patch segmentation for human full–body and body–parts detection and tracking. IEEE: Proceedings of AVSS. https://doi.org/10.1109/AVSS.2014.6918695
DAS S, Chuadhary A, Bremond F, Thonnat M (2019) Where to focus on for human action recognition?. IEEE Winter Conference on Applications of computer vision. https://doi.org/10.1109/WACV.2019.00015
Daniel W, Remi R, Edmond B (2006) Free Viewpoint Action Recognition using Motion History Volumes. Comput Vis Image Underst (CVIU) 104:249–257
Dargazany A, Nicolescu M (2012) Human body parts tracking using torso tracking: applications to activity recognition. Proceedings of ITNG. arXiv:1907.05281
Dawn DD, Shaikh SH (2016) A comprehensive survey of human action recognition with spatio–temporal interest point (STIP) detector. The Vis Comput 32:289–306
Fei M, Ju Z, Zhen X, Li J (2017) Real-time Visual Tracking based on Improved Perceptual Hashing [J]. Multimed Tools Appl 76(3):4617–4634
Guo Y, Yue X, Yan G (2013) Salient region detection based on multi–resolution. IEEE: International Conference on Machine learning and Cybernetics. https://doi.org/10.1109/ICMLC.2013.6890422
Han Y, Chung S, Ambikapathi A, Chan J, Lin W, Su S (2018) Robust human action recognition using global spatial–temporal attention for human skeleton data. Proceedings of IJCNN. DOI: https://doi.org/10.1109/IJCNN.2018.8489386
Hu Z, Lin X, Yan H (2006) Torso Detection in Static Images. IEEE: International Conference on Signal. https://doi.org/10.1109/ICOSP.2006.345837. Processing
Hussain I (2019) AAMAZ Human Action Recognition Dataset, Kaggle
Ignatov A (2018) Real–time human activity recognition from accelerometer data using Convolutional Neural Networks. Appl Soft Comput 62:915–922
Jalal A, Zia-Uddin M, Kim T (2012) Depth Video–based Human Activity Recognition System Using Translation and Scaling Invariant Features for Life Logging at Smart Home, IEEE Transaction on Consumer Electronics, ISSN: 0098–3063 58(3):863–871
Jalal A, Sharif N, Kim J et al (2013) Human activity recognition via recognized body parts of human depth silhouettes for residents monitoring services at smart homes. Indoor Built Environ 22:271–279
Jalal A, Kamal S, Kim D (2014) A depth video sensor–based life–logging human activity recognition system for elderly care in smart indoor environments. Sensors 14(7):11735–11759
Jalal A, Kim Y, Kim D (2014) Ridge body parts features for human pose estimation and recognition from RGB–D video data. Proceedings of the IEEE International Conference on computing, communication and networking technologies, pp. 1–6
Jalal A, Mahmood M, Sidduqi M (2018) Robust spatio–temporal features for human interaction recognition via artificial neural network, IEEE conference on International Conference on Frontiers of information technology
Jalal A, Nadeem A, Bobasu S (2019) Human body parts estimation and detection for physical sports movements. IEEE International Conference on Communication, Computing and Digital Systems
Jalal A, Khalid N, Kim K (2020) Automatic Recognition of Human Interaction via Hybrid Descriptors and Maximum Entropy Markov Model Using Depth Sensors. Entropy
Jaouedi N, Boujnah N, Bouhlel MS (2019) A new hybrid deep learning model for human action recognition. Journal of King Saud University – Computer and Information Sciences. https://doi.org/10.1016/j.jksuci.2019.09.004
Johnson E (2010) Clustered Pose and Non–linear Appearance Models for Human Pose Estimation. Proceedings of the British Machine Vision Conference. https://doi.org/10.5244/C.24.12
Kim Y, Kim D Real-time dance evaluation by markerless human pose estimation. Multimed Tools Appl. https://doi.org/10.1007/s11042-018-6068-4
Khan MUS, Abbas A, Ali M (2018) On the Correlation of Sensor Location and Human Activity Recognition in Body Area Networks (BANs). IEEE Syst J 12:82–91
Li J, Li X, Tao D (2008) KPCA for Semantic Object Extraction in Images [J] Pattern Recognition 41(10):3244–3250
Li N, Wen L, Dong X (2015) Visual recognition by learning from web data: A weakly supervised domain generalization approach. IEEE Conf. Comput. Vis. Pattern Recognit (CVPR) https://doi.org/10.1109/CVPR.2015.7298894
Li G, Tang H, Sun Y et al (2019) Hand gesture recognition based on convolution neural network. Cluster Comput 22(Supplement 2): 2719–2729. https://doi.org/10.1007/s10586-017-1435-x
Li C, Zhang B, Chen C et al (2019) Deep Manifold Structure Transfer for Action Recognition. IEEE transactions on image processing 28(9) https://doi.org/10.1109/TIP.2019.2912357
Liu M, Liu H, Chen C (2017) Enhanced skeleton visualization for view invariant human action recognition. Pattern Recogn 68:346–362
Liu C, Yuen PC (2011) A Boosted Co–Training Algorithm for Human Action Recognition. IEEE Trans Circ Syst Video Technol 21:1203–1213
Liu M, Liu H, Sun Q, Zhang T, Ding R (2016) Salient pairwise spatio–temporal interest points for real–time activity recognition. CAAI Trans Intell Technol 1:14–29
Liu T, Stathaki T (2016) Fast head–shoulder proposal for deformable part model based pedestrian detection. IEEE International Conference on Digital Signal Processing (DSP). https://doi.org/10.1109/ICDSP.2016.7868599
Liu F, Xu X, Qiu S, Qing C, Tao D (2016) Simple to complex transfer learning for action recognition. IEEE Trans Image Process 25:949–960
Luvizon DC, Hedi T, David P (2017) Learning features combination for human action recognition from skeleton sequences. Pattern Recogn Lett 99:13–20
Madabhushi A, Aggarwal J (1999) A bayesian approach to human activity recognition, IEEE Visual Surveillance. https://doi.org/10.1109/VS.1999.780265
Mahmood M, Jalal A, Kim K (2020) WHITE STAG Model: Wise Human Interaction Tracking and Estimation (WHITE) using Spatio–temporal and Angular–geometric (STAG) Descriptors, Multimedia Tools and Applications
Manzi A, Moschetti A, Limosani R, Fiorini L, Cavallo F (2018) Enhancing Activity Recognition of Self–Localized Robot Through Depth Camera and Wearable Sensors. IEEE Sens J 18:9324–9331
Milanova M, Ali S, Al-Rizzo H, Fox VL (2015) Human action Recognition: Contour–based and silhouette–based Approaches. Springer Cham. https://doi.org/10.1007/978-3-319-11430-9
Mojarrad M, Dezfouli M, Rahmani A (2008) Feature’s Extraction of Human Body Composition in Images by Segmentation Method. Pwaset 35:267–270
Nadeem A, Jalal A, Kim K (2020) Human actions tracking and recognition based on body parts detection via Artificial neural network. IEEE International Conference on Advancements in computational sciences
Nguyen ND, Bui DT, Truong PH, Jeong GM (2018) Classification of Five Ambulatory Activities Regarding Stair and Incline Walking Using Smart Shoes. IEEE Sensors Journal. https://doi.org/10.1109/JSEN.2018.2837674
Osterland S, Weber J (2019) Analytical analysis of single–stage pressure relief valves. Int J Hydromechatron 2(1):32–53
Quaid M, Jalal A (2019) Wearable Sensors based Human Behavioral Pattern Recognition using Statistical Features and Reweighted Genetic Algorithm. Multimedia Tools and Applications
Rezaie H, Ghassemian M (2017) An Adaptive Algorithm to Improve Energy Efficiency in Wearable Activity Recognition Systems. IEEE Sens J 17:5315–5323
Riemenschneider H, Donoser M, Bischof H (2009) Bag of Optical Flow Volumes for Image Sequence Recognition. British Machine Vision Conference. https://doi.org/10.5244/C.23.28
Sadanand S, Corso JJ (2012) Action bank: A high–level representation of activity invideo. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit, https://doi.org/10.1109/CVPR.2012.6247806
San-Segundo R, Blunck H, Moreno-Pimentel J, Stisen A, Gil-Martín M (2018) Robust Human Activity Recognition using smartwatches and smartphones. Eng Appl Artif Intell 72:190–202
Shebiah RN, Sangari AA (2019) Classification of human body parts using histogram of oriented gradients. Proceedings of ICACCS. https://doi.org/10.1109/ICACCS.2019.8728328
Shokri M, Tavakoli K (2019) A review on the artificial neural network approach to analysis and prediction of seismic damage in infrastructure. Int J Hydromechatron 2(4):178–196
Sun Y et al (2020) Intelligent human computer interaction based on non redundant EMG signal. Alexandria Engineering Journal https://doi.org/10.1016/j.aej.2020.01.015
Susan S, Agrawal P, Mittal M et al (2019) New shape descriptor in the context of edge continuity, CAAI Transactions on Intelligence Technology 4(2)
Tingting Y, Junqian W, Lintai W et al (2019) Three–stage network for age estimation, CAAI Transactions on Intelligence Technology 4(2)
Vig E, Dorr M, Cox D (2012) Space–variant descriptor sampling for action recognition based on saliency and eye movements. European Conference of Computer Vision. https://doi.org/10.1007/978--3--642--33786--4
Wang Y, Cang S, Yu H (2018) A Data Fusion–Based Hybrid Sensory System for Older People’s Daily Activity and Daily Routine Recognition. IEEE Sens J 18:6874–6888
Wiens T (2019) Engine speed reduction for hydraulic machinery using predictive algorithms. Int J Hydromechatron 2(1):16–31
Xia L, Chen CC, Aggarwal JK (2012) View invariant human action recognition using histograms of 3D joints. Proceedings of CVPRW. https://doi.org/10.1109/CVPRW.2012.6239233
Xie C, Li C, Zhang B et al Memory Attention Networks for Skeleton-based Action Recognition. arXiv:https://arxiv.org/abs/1804.08254v2
Yue H, Chen W (2015) Comments on Automatic Visual Bag–of–Words for Online Robot Navigation and Mapping. IEEE Transactions on Robotics 31:223–224
Zhang B, Yang Y, Chen et al (2017) Action Recognition Using 3D Histograms of Texture and A Multi–Class Boosting Classifier. IEEE transactions on image processing 26(10). https://doi.org/10.1109/TIP.2017.2718189
Zhang J, Shum H, Han J et al (2018) Action Recognition From Arbitrary Views Using Transferable Dictionary Learning. IEEE transactions on image processing 27(10). https://doi.org/10.1109/TIP.2018.2836323
Zhu C, Miao D (2019) Influence of kernel clustering on an RBFN, CAAI Transactions on Intelligence Technology 4(4)
Acknowledgements
This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No. 2018R1D1A1A02085645) and by a grant (19CTAP-C152247-01) from Technology Advancement Research Program funded by Ministry of Land, Infrastructure and Transport of Korean government.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Nadeem, A., Jalal, A. & Kim, K. Automatic human posture estimation for sport activity recognition with robust body parts detection and entropy markov model. Multimed Tools Appl 80, 21465–21498 (2021). https://doi.org/10.1007/s11042-021-10687-5
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-021-10687-5