Back to Search Start Over

Deep Learning for Human Activity Recognition on 3D Human Skeleton: Survey and Comparative Study.

Authors :
Nguyen, Hung-Cuong
Nguyen, Thi-Hao
Scherer, Rafał
Le, Van-Hung
Source :
Sensors (14248220). Jun2023, Vol. 23 Issue 11, p5121. 33p.
Publication Year :
2023

Abstract

Human activity recognition (HAR) is an important research problem in computer vision. This problem is widely applied to building applications in human–machine interactions, monitoring, etc. Especially, HAR based on the human skeleton creates intuitive applications. Therefore, determining the current results of these studies is very important in selecting solutions and developing commercial products. In this paper, we perform a full survey on using deep learning to recognize human activity based on three-dimensional (3D) human skeleton data as input. Our research is based on four types of deep learning networks for activity recognition based on extracted feature vectors: Recurrent Neural Network (RNN) using extracted activity sequence features; Convolutional Neural Network (CNN) uses feature vectors extracted based on the projection of the skeleton into the image space; Graph Convolution Network (GCN) uses features extracted from the skeleton graph and the temporal–spatial function of the skeleton; Hybrid Deep Neural Network (Hybrid–DNN) uses many other types of features in combination. Our survey research is fully implemented from models, databases, metrics, and results from 2019 to March 2023, and they are presented in ascending order of time. In particular, we also carried out a comparative study on HAR based on a 3D human skeleton on the KLHA3D 102 and KLYOGA3D datasets. At the same time, we performed analysis and discussed the obtained results when applying CNN-based, GCN-based, and Hybrid–DNN-based deep learning networks. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
14248220
Volume :
23
Issue :
11
Database :
Academic Search Index
Journal :
Sensors (14248220)
Publication Type :
Academic Journal
Accession number :
164216760
Full Text :
https://doi.org/10.3390/s23115121