A Deep Learning Approach for Human Action Recognition Using Skeletal Information.

Journal: Advances in experimental medicine and biology
Published Date:

Abstract

In this paper we present an approach toward human action detection for activities of daily living (ADLs) that uses a convolutional neural network (CNN). The network is trained on discrete Fourier transform (DFT) images that result from raw sensor readings, i.e., each human action is ultimately described by an image. More specifically, we work using 3D skeletal positions of human joints, which originate from processing of raw RGB sequences enhanced by depth information. The motion of each joint may be described by a combination of three 1D signals, representing its coefficients into a 3D Euclidean space. All such signals from a set of human joints are concatenated to form an image, which is then transformed by DFT and is used for training and evaluation of a CNN. We evaluate our approach using a publicly available challenging dataset of human actions that may involve one or more body parts simultaneously and for two sets of actions which resemble common ADLs.

Authors

  • Eirini Mathe
    Institute of Informatics and Telecommunications, National Center for Scientific Research- "Demokritos", Athens, Greece.
  • Apostolos Maniatis
    Department of Computer Engineering T.E, Technological Education Institute of Sterea Ellada, Lamia, Greece.
  • Evaggelos Spyrou
    Institute of Informatics and Telecommunications, National Center for Scientific Research-"Demokritos", Athens, Greece. espyrou@iit.demokritos.gr.
  • Phivos Mylonas
    Department of Informatics, Ionian University, Corfu, Greece. fmylonas@ionio.gr.