Action Recognition with Skeletal Volume and Deep Learning


KEÇELİ A. S. , KAYA A., CAN A. B.

25th Signal Processing and Communications Applications Conference (SIU), Antalya, Turkey, 15 - 18 May 2017 identifier identifier

  • Publication Type: Conference Paper / Full Text
  • Doi Number: 10.1109/siu.2017.7960280
  • City: Antalya
  • Country: Turkey
  • Keywords: Action recognition, RGBD data, deep learning, SVM, feature selection, POSE

Abstract

The use of depth sensors in activity recognition is a technology that emerges in human computer interaction and motion recognition. In this study, an approach to identify single-person activities using deep learning on depth image sequences is presented. First, a 3D volumetric template is generated using skeletal information obtained from a depth video. The generated 3D volume is used for extracting features by taking images from different angles at different volumes. Actions are recognized by extracting deep features using AlexNet model [1] and Histogram of Oriented Gradients (HOG) features from these images. The proposed method has been tested with MSRAction3D [2] and UTHKinect-Action3D [2] datasets. The obtained results were comparable to similar studies in the literature.