Sorry, I don't understand your search. ×
Back to Search Start Over

Multi-Modality Multi-Task Recurrent Neural Network for Online Action Detection

Authors :
Yanghao Li
Junliang Xing
Jiaying Liu
Wenjun Zeng
Cuiling Lan
Sijie Song
Source :
IEEE Transactions on Circuits and Systems for Video Technology. 29:2667-2682
Publication Year :
2019
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2019.

Abstract

Online action detection is a brand new challenge and plays a critical role in visual surveillance analytics. It goes one step further than a conventional action recognition task, which recognizes human actions from well-segmented clips. Online action detection is desired to identify the action type and localize action positions on the fly from the untrimmed stream data. In this paper, we propose a multi-modality multi-task recurrent neural network, which incorporates both RGB and Skeleton networks. We design different temporal modeling networks to capture specific characteristics from various modalities. Then, a deep long short-term memory subnetwork is utilized effectively to capture the complex long-range temporal dynamics, naturally avoiding the conventional sliding window design and thus ensuring high computational efficiency. Constrained by a multi-task objective function in the training phase, this network achieves superior detection performance and is capable of automatically localizing the start and end points of actions more accurately. Furthermore, embedding subtask of regression provides the ability to forecast the action prior to its occurrence. We evaluate the proposed method and several other methods in action detection and forecasting on the online action detection data set and gaming action data set datasets. Experimental results demonstrate that our model achieves the state-of-the-art performance on both tasks.

Details

ISSN :
15582205 and 10518215
Volume :
29
Database :
OpenAIRE
Journal :
IEEE Transactions on Circuits and Systems for Video Technology
Accession number :
edsair.doi...........ba7bcb9e518a1ab6576cb0a9a454f24d
Full Text :
https://doi.org/10.1109/tcsvt.2018.2799968