Item request has been placed! ×
Item request cannot be made. ×
loading  Processing Request

Spatio-Temporal Image Representation of 3D Skeletal Movements for View-Invariant Action Recognition with Deep Convolutional Neural Networks

Item request has been placed! ×
Item request cannot be made. ×
loading   Processing Request
  • معلومة اضافية
    • Contributors:
      European Commission; Ministerio de Economía y Competitividad (España)
    • بيانات النشر:
      MDPI
    • الموضوع:
      2019
    • Collection:
      Universidad Carlos III de Madrid: e-Archivo
    • نبذة مختصرة :
      This article belongs to the Special Issue Deep Learning-Based Image Sensors ; Designing motion representations for 3D human action recognition from skeleton sequences is an important yet challenging task. An effective representation should be robust to noise, invariant to viewpoint changes and result in a good performance with low-computational demand. Two main challenges in this task include how to efficiently represent spatio-temporal patterns of skeletal movements and how to learn their discriminative features for classification tasks. This paper presents a novel skeleton-based representation and a deep learning framework for 3D action recognition using RGB-D sensors. We propose to build an action map called SPMF (Skeleton Posture-Motion Feature), which is a compact image representation built from skeleton poses and their motions. An Adaptive Histogram Equalization (AHE) algorithm is then applied on the SPMF to enhance their local patterns and form an enhanced action map, namely Enhanced-SPMF. For learning and classification tasks, we exploit Deep Convolutional Neural Networks based on the DenseNet architecture to learn directly an end-to-end mapping between input skeleton sequences and their action labels via the Enhanced-SPMFs. The proposed method is evaluated on four challenging benchmark datasets, including both individual actions, interactions, multiview and large-scale datasets. The experimental results demonstrate that the proposed method outperforms previous state-of-the-art approaches on all benchmark tasks, whilst requiring low computational time for training and inference. ; This research was carried out at the Cerema Research Center and Informatics Research Institute of Toulouse, Paul Sabatier University, France. The authors would like to express our thanks to all the people who have made helpful comments and suggestions on a previous draft. S.A. Velastin is grateful to funding received from the Universidad Carlos III de Madrid, the European Union’s Seventh Framework Programme for research, ...
    • ISSN:
      1424-8220
    • Relation:
      info:eu-repo/grantAgreement/EC/H2020/600371; Gobierno de España. COFUND2013-51509; Pham, H.H., Salmane, H., Khoudour, L., Crouzil, A., Zegers, P. y Velastín, S. (2019). Spatio–Temporal Image Representation of 3D Skeletal Movements for View-Invariant Action Recognition with Deep Convolutional Neural Networks. Sensors, 19(8), 1932.; http://hdl.handle.net/10016/28852; https://doi.org/10.3390/s19081932; Sensors; 19; AR/0000023773
    • الرقم المعرف:
      10.3390/s19081932
    • الدخول الالكتروني :
      http://hdl.handle.net/10016/28852
      https://doi.org/10.3390/s19081932
    • Rights:
      © 2019 The Authors. ; Atribución 3.0 España ; http://creativecommons.org/licenses/by/3.0/es/ ; open access
    • الرقم المعرف:
      edsbas.7B09541A