Back to Search
Start Over
Recognizing human actions by learning and matching shape-motion prototype trees.
- Source :
-
IEEE transactions on pattern analysis and machine intelligence [IEEE Trans Pattern Anal Mach Intell] 2012 Mar; Vol. 34 (3), pp. 533-47. - Publication Year :
- 2012
-
Abstract
- A shape-motion prototype-based approach is introduced for action recognition. The approach represents an action as a sequence of prototypes for efficient and flexible action matching in long video sequences. During training, an action prototype tree is learned in a joint shape and motion space via hierarchical K-means clustering and each training sequence is represented as a labeled prototype sequence; then a look-up table of prototype-to-prototype distances is generated. During testing, based on a joint probability model of the actor location and action prototype, the actor is tracked while a frame-to-prototype correspondence is established by maximizing the joint probability, which is efficiently performed by searching the learned prototype tree; then actions are recognized using dynamic prototype sequence matching. Distance measures used for sequence matching are rapidly obtained by look-up table indexing, which is an order of magnitude faster than brute-force computation of frame-to-frame distances. Our approach enables robust action matching in challenging situations (such as moving cameras, dynamic backgrounds) and allows automatic alignment of action sequences. Experimental results demonstrate that our approach achieves recognition rates of 92.86 percent on a large gesture data set (with dynamic backgrounds), 100 percent on the Weizmann action data set, 95.77 percent on the KTH action data set, 88 percent on the UCF sports data set, and 87.27 percent on the CMU action data set.
Details
- Language :
- English
- ISSN :
- 1939-3539
- Volume :
- 34
- Issue :
- 3
- Database :
- MEDLINE
- Journal :
- IEEE transactions on pattern analysis and machine intelligence
- Publication Type :
- Academic Journal
- Accession number :
- 21788666
- Full Text :
- https://doi.org/10.1109/TPAMI.2011.147