Deep-Temporal LSTM for Daily Living Action Recognition - Archive ouverte HAL Access content directly
Conference Papers Year :

Deep-Temporal LSTM for Daily Living Action Recognition

(1) , (1) , (1) , (2)
1
2
Srijan Das
  • Function : Author
  • PersonId : 21855
  • IdHAL : srijan-das
Michal F Koperski
Gianpiero Francesca
  • Function : Author
  • PersonId : 1023188

Abstract

In this paper, we propose to improve the traditional use of RNNs by employing a many to many model for video classification. We analyze the importance of modeling spatial layout and temporal encoding for daily living action recognition. Many RGB methods focus only on short term temporal information obtained from optical flow. Skeleton based methods on the other hand show that modeling long term skeleton evolution improves action recognition accuracy. In this work, we propose a deep-temporal LSTM architecture which extends standard LSTM and allows better encoding of temporal information. In addition, we propose to fuse 3D skeleton geometry with deep static appearance. We validate our approach on public available CAD60, MSRDai-lyActivity3D and NTU-RGB+D, achieving competitive performance as compared to the state-of-the art.
Fichier principal
Vignette du fichier
avss-2018.pdf (466.78 Ko) Télécharger le fichier
Origin : Files produced by the author(s)
Loading...

Dates and versions

hal-01896064 , version 1 (15-10-2018)

Identifiers

  • HAL Id : hal-01896064 , version 1

Cite

Srijan Das, Michal F Koperski, Francois Bremond, Gianpiero Francesca. Deep-Temporal LSTM for Daily Living Action Recognition. 15th IEEE International Conference on Advanced Video and Signal-based Surveillance, Nov 2018, Auckland, New Zealand. ⟨hal-01896064⟩
67 View
241 Download

Share

Gmail Facebook Twitter LinkedIn More