An Efficient Human Action Recognition Framework with Pose-Based Spatiotemporal Features
Loading...

Date
2020
Authors
Agahian, Saeid
Negin, Farhood
Kose, Cemal
Journal Title
Journal ISSN
Volume Title
Publisher
Elsevier - Division Reed Elsevier India Pvt Ltd
Open Access Color
OpenAIRE Downloads
OpenAIRE Views
Abstract
In the past two decades, human action recognition has been among the most challenging tasks in the field of computer vision. Recently, extracting accurate and cost-efficient skeleton information became available thanks to the cutting edge deep learning algorithms and low-cost depth sensors. In this paper, we propose a novel framework to recognize human actions using 3D skeleton information. The main components of the framework are pose representation and encoding. Assuming that human actions can be represented by spatiotemporal poses, we define a pose descriptor consisting of three elements. The first element contains the normalized coordinates of the raw skeleton joints information. The second element contains the temporal displacement information relative to a predefined temporal offset and the third element keeps the displacement information pertinent to the previous timestamp in the temporal resolution. The final descriptor of the whole sequence is the concatenation of frame-wise descriptors. To avoid the problems regarding high dimensionality, Principal Component Analysis (PCA) is applied on the descriptors. The resulted descriptors are encoded with Fisher Vector (FV) representation before they get trained with an Extreme Learning Machine (ELM). The performance of the proposed framework is evaluated by three public benchmark datasets. The proposed method achieved competitive results compared to the other methods in the literature. (C) 2019 Karabuk University. Publishing services by Elsevier B.V.
Description
Alp, Sait/0000-0003-2462-6166;
ORCID
Keywords
Skeleton-Based, 3D Action Recognition, Extreme Learning Machines, RGB-D
Fields of Science
Citation
WoS Q
Q1
Scopus Q
Q1
Source
Engineering Science and Technology-An International Journal-JESTECH
Volume
23
Issue
1
Start Page
196
End Page
203
