TY - JOUR
T1 - Enhancing Surveillance Vision-Based Human Action Recognition Using Skeleton Joint Swing and Angle Feature and Modified AlexNet-LSTM
AU - Yunardi, Riky Tri
AU - Sardjono, Tri Arief
AU - Mardiyanto, Ronny
N1 - Publisher Copyright:
© (2025), (Intelligent Network and Systems Society). All rights reserved.
PY - 2025
Y1 - 2025
N2 - Human action recognition (HAR) identifies and classifies human activities by analyzing activity patterns, with feature extraction and classification accuracy being key challenges. In the field of surveillance vision-based, it requires the ability to accurately detect suspicious human activities to provide public safety. In the classifier model for action recognition, the steps start from feature extraction to classification. The classification step uses recurrent neural network architectures such as LSTM to handle sequential data. However, this approach struggles to process spatial information in video data, necessitating the need for a model to learn spatiotemporal patterns from feature data. To address these issues, this study proposes a novel method for classifying activities based on the pattern of 2D skeleton joint swing and angle features for each activity. Additionally, it introduces a novel modified AlexNet architecture with two LSTM layers, called AlexNet-2LSTM, to improve the accuracy of human activity classification. In the performance experiments, the proposed method was evaluated on the KTH and Weizmann datasets, both of which include videos of several people performing different actions. Moreover, to demonstrate the accuracy of the proposed classifier model, it was compared against other state-of-the-art (SOTA) deep learning classifiers, namely Optimized-LSTM, Triple Parallel LSTM, Hybrid CNN-LSTM, LCSWnet, and CC-LSTM-CNN, which the AlexNet-2LSTM achieved precision of 0.95, recall of 0.95, F1-score of 0.94, and accuracy of 0.96on the KTH dataset. Besides that, on the Weizmann dataset 0.95, 0.94, 0.94, and 0.93, respectively. These achievements highlight the proposed model contribution to improving feature extraction and classification accuracy in vision-based HAR systems.
AB - Human action recognition (HAR) identifies and classifies human activities by analyzing activity patterns, with feature extraction and classification accuracy being key challenges. In the field of surveillance vision-based, it requires the ability to accurately detect suspicious human activities to provide public safety. In the classifier model for action recognition, the steps start from feature extraction to classification. The classification step uses recurrent neural network architectures such as LSTM to handle sequential data. However, this approach struggles to process spatial information in video data, necessitating the need for a model to learn spatiotemporal patterns from feature data. To address these issues, this study proposes a novel method for classifying activities based on the pattern of 2D skeleton joint swing and angle features for each activity. Additionally, it introduces a novel modified AlexNet architecture with two LSTM layers, called AlexNet-2LSTM, to improve the accuracy of human activity classification. In the performance experiments, the proposed method was evaluated on the KTH and Weizmann datasets, both of which include videos of several people performing different actions. Moreover, to demonstrate the accuracy of the proposed classifier model, it was compared against other state-of-the-art (SOTA) deep learning classifiers, namely Optimized-LSTM, Triple Parallel LSTM, Hybrid CNN-LSTM, LCSWnet, and CC-LSTM-CNN, which the AlexNet-2LSTM achieved precision of 0.95, recall of 0.95, F1-score of 0.94, and accuracy of 0.96on the KTH dataset. Besides that, on the Weizmann dataset 0.95, 0.94, 0.94, and 0.93, respectively. These achievements highlight the proposed model contribution to improving feature extraction and classification accuracy in vision-based HAR systems.
KW - Deep learning classifier
KW - Long short-term memory
KW - Skeleton joint
KW - Vision-based human action recognition
UR - http://www.scopus.com/inward/record.url?scp=85214246687&partnerID=8YFLogxK
U2 - 10.22266/ijies2025.0229.53
DO - 10.22266/ijies2025.0229.53
M3 - Article
AN - SCOPUS:85214246687
SN - 2185-310X
VL - 18
SP - 754
EP - 768
JO - International Journal of Intelligent Engineering and Systems
JF - International Journal of Intelligent Engineering and Systems
IS - 1
ER -