Recurrent Neural Network for Human Action Recognition using Star Skeletonization
DOI:
https://doi.org/10.32628/CSEIT195217Keywords:
Human Action Recognition, Video Streams, RNN, Star skeletonization.Abstract
Human Action Recognition has been an active research topic since early 1980s due to its promising applications in many domains like video indexing, surveillance, gesture recognition, video retrieval and human-computer interactions where the actions in the form of videos or sensor datas are recognized. The extraction of relevant features from the video streams is the most challenging part. With the emergence of advanced artificial intelligence techniques, deep learning methods are adopted to achieve the goal. The proposed system presents a Recurrent Neural Network (RNN) methodology for Human Action Recognition using star skeleton as a representative descriptor of human posture. Star skeleton is the process of jointing the gross contour extremes of a body to its centroid. To use star skeleton as feature for action recognition, the feature is defined as a five-dimensional vector in star fashion because the head and four limbs are usually local extremes of human body. In our project, we assumed an action is composed of a series of star skeletons overtime. Therefore, images expressing human action which are time-sequential are transformed into a feature vector sequence. Then the feature vector sequence must be transformed into symbol sequence so that RNN can model the action. RNN is used because the features extracted are time dependent
References
- Aouaidjia Kamel, Bin Sheng ,Bin Sheng,Po Yang, Ping Li,Ruimin Shen, and David Dagan Feng, “Deep Convolutional Neural Networks for Human Action Recognition Using Depth Maps and Postures”, IEEE transactions on system,man and cybernetics systems.
- Tapia, E.M., Intille, S.S., Larson, K.: Activity recognition in the home using simple and ubiquitous sensors. In: Ferscha, A., Mattern, F. (eds.) Pervasive 2004. LNCS, vol. 3001, pp. 158–175. Springer, Heidelberg (2004). doi:10.1007/ 978-3-540-24646-6 10
- Bao, L., Intille, S.S.: Activity recognition from user-annotated acceleration data. In: Ferscha, A., Mattern, F. (eds.) Pervasive 2004. LNCS, vol. 3001, pp. 1–17. Springer, Heidelberg (2004). doi:10.1007/978-3-540-24646-6 1 14.
- Van Kasteren, T., Noulas, A., Englebienne, G., Kr¨ose, B.: Accurate activity recognition in a home setting. In: Proceedings of the 10th International Conference on Ubiquitous Computing, pp. 1–9. ACM (2008).
- Wu, W., Dasgupta, S., Ramirez, E.E., Peterson, C., Norman, G.J.: Classification accuracies of physical activities using smartphone motion sensors. J. Med. Internet Res. 14, e130 (2012).
- Zhu, Y., Nayak, N.M., Roy-Chowdhury, A.K.: Context-aware activity recognition and anomaly detection in video. IEEE J. Sel. Top. Sig. Proces. 7, 91–101 (2013).
- Duckworth, M. Alomari, Y. Gatsoulis, D. C. Hogg, and A. G. Cohn, “Unsupervised activity recognition using latent semantic analysis on a mobile robot,” in Proc. Eur. Conf.Artif. Intell., 2016, pp. 1062–1070.
- M. H. Kolekar and D. P. Dash, “Hidden Markov model based human activity recognition using shape and optical flow based features,” in Proc. IEEE Region 10 Conf.,2016, pp. 393–397.
- G. Liang, X. Lan, J. Wang, J. Wang, and N. Zheng, “A limb-based graphical model for human pose estimation,” IEEE Trans. Syst., Man, Cybern., Syst., vol. 48, no. 7, pp.1080–1092, Jul. 2018
- J. Sriwan and W. Suntiamorntut, “Human activity monitoring system based on WSNs,” in Proc. Int. Joint Conf. Comput. Sci. Softw. Eng., 2015, pp. 247–250.
- Y. Guo, D. Tao, W. Liu, and J. Cheng, “Multiview Cauchy estimator feature embedding
Downloads
Published
Issue
Section
License
Copyright (c) IJSRCSEIT

This work is licensed under a Creative Commons Attribution 4.0 International License.