Literature DB >> 23681992

Discovering motion primitives for unsupervised grouping and one-shot learning of human actions, gestures, and expressions.

Yang Yang1, Imran Saleemi, Mubarak Shah.   

Abstract

This paper proposes a novel representation of articulated human actions and gestures and facial expressions. The main goals of the proposed approach are: 1) to enable recognition using very few examples, i.e., one or k-shot learning, and 2) meaningful organization of unlabeled datasets by unsupervised clustering. Our proposed representation is obtained by automatically discovering high-level subactions or motion primitives, by hierarchical clustering of observed optical flow in four-dimensional, spatial, and motion flow space. The completely unsupervised proposed method, in contrast to state-of-the-art representations like bag of video words, provides a meaningful representation conducive to visual interpretation and textual labeling. Each primitive action depicts an atomic subaction, like directional motion of limb or torso, and is represented by a mixture of four-dimensional Gaussian distributions. For one--shot and k-shot learning, the sequence of primitive labels discovered in a test video are labeled using KL divergence, and can then be represented as a string and matched against similar strings of training videos. The same sequence can also be collapsed into a histogram of primitives or be used to learn a Hidden Markov model to represent classes. We have performed extensive experiments on recognition by one and k-shot learning as well as unsupervised action clustering on six human actions and gesture datasets, a composite dataset, and a database of facial expressions. These experiments confirm the validity and discriminative nature of the proposed representation.

Entities:  

Mesh:

Year:  2013        PMID: 23681992     DOI: 10.1109/TPAMI.2012.253

Source DB:  PubMed          Journal:  IEEE Trans Pattern Anal Mach Intell        ISSN: 0098-5589            Impact factor:   6.226


  6 in total

1.  Discovery and recognition of motion primitives in human activities.

Authors:  Marta Sanzari; Valsamis Ntouskos; Fiora Pirri
Journal:  PLoS One       Date:  2019-04-01       Impact factor: 3.240

2.  A Branch-and-Bound Framework for Unsupervised Common Event Discovery.

Authors:  Wen-Sheng Chu; Fernando De la Torre; Jeffrey F Cohn; Daniel S Messinger
Journal:  Int J Comput Vis       Date:  2017-02-09       Impact factor: 7.410

3.  Human Pose Estimation from Monocular Images: A Comprehensive Survey.

Authors:  Wenjuan Gong; Xuena Zhang; Jordi Gonzàlez; Andrews Sobral; Thierry Bouwmans; Changhe Tu; El-Hadi Zahzah
Journal:  Sensors (Basel)       Date:  2016-11-25       Impact factor: 3.576

4.  An Unsupervised Framework for Online Spatiotemporal Detection of Activities of Daily Living by Hierarchical Activity Models.

Authors:  Farhood Negin; François Brémond
Journal:  Sensors (Basel)       Date:  2019-09-29       Impact factor: 3.576

5.  Toward Modeling Psychomotor Performance in Karate Combats Using Computer Vision Pose Estimation.

Authors:  Jon Echeverria; Olga C Santos
Journal:  Sensors (Basel)       Date:  2021-12-15       Impact factor: 3.576

6.  Behavior Discovery and Alignment of Articulated Object Classes from Unstructured Video.

Authors:  Luca Del Pero; Susanna Ricco; Rahul Sukthankar; Vittorio Ferrari
Journal:  Int J Comput Vis       Date:  2016-08-10       Impact factor: 7.410

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.