Pose-conditioned Spatio-Temporal Attention for Human Action Recognition

Bastien Moysset 1 Christoper Kermorvant Christian Wolf 1
1 imagine - Extraction de Caractéristiques et Identification
LIRIS - Laboratoire d'InfoRmatique en Image et Systèmes d'information
Abstract : We address human action recognition from multi-modal video data involving articulated pose and RGB frames and propose a two-stream approach. The pose stream is processed with a convolutional model taking as input a 3D tensor holding data from a sub-sequence. A specific joint ordering, which respects the topology of the human body, ensures that different convolutional layers correspond to meaningful levels of abstraction. The raw RGB stream is handled by a spatio-temporal soft-attention mechanism conditioned on features from the pose network. An LSTM network receives input from a set of image locations at each instant. A trainable glimpse sensor extracts features on a set of predefined locations specified by the pose stream, namely the 4 hands of the two people involved in the activity. Appearance features give important cues on hand motion and on objects held in each hand. We show that it is of high interest to shift the attention to different hands at different time steps depending on the activity itself. Finally a temporal attention mechanism learns how to fuse LSTM features over time. We evaluate the method on 3 datasets. State-of-the-art results are achieved on the largest dataset for human activity recognition, namely NTU-RGB+D, as well as on the SBU Kinect Interaction dataset. Performance close to state-of-the-art is achieved on the smaller MSR Daily Activity 3D dataset.
Complete list of metadatas

Contributor : Christian Wolf <>
Submitted on : Tuesday, September 26, 2017 - 2:32:01 PM
Last modification on : Tuesday, February 26, 2019 - 4:35:36 PM

Links full text


  • HAL Id : hal-01593572, version 1
  • ARXIV : 1703.10106


Bastien Moysset, Christoper Kermorvant, Christian Wolf. Pose-conditioned Spatio-Temporal Attention for Human Action Recognition. 2017. ⟨hal-01593572⟩



Record views