3D pose from motion for cross-view action recognition via non-linear circulant temporal encoding

A Gupta, J Martinez, JJ Little… - Proceedings of the …, 2014 - openaccess.thecvf.com
Proceedings of the IEEE conference on computer vision and …, 2014openaccess.thecvf.com
We describe a new approach to transfer knowledge across views for action recognition by
using examples from a large collection of unlabelled mocap data. We achieve this by directly
matching purely motion based features from videos to mocap. Our approach recovers 3D
pose sequences without performing any body part tracking. We use these matches to
generate multiple motion projections and thus add view invariance to our action recognition
model. We also introduce a closed form solution for approximate non-linear Circulant …
Abstract
We describe a new approach to transfer knowledge across views for action recognition by using examples from a large collection of unlabelled mocap data. We achieve this by directly matching purely motion based features from videos to mocap. Our approach recovers 3D pose sequences without performing any body part tracking. We use these matches to generate multiple motion projections and thus add view invariance to our action recognition model. We also introduce a closed form solution for approximate non-linear Circulant Temporal Encoding (nCTE), which allows us to efficiently perform the matches in the frequency domain. We test our approach on the challenging unsupervised modality of the IXMAS dataset, and use publicly available motion capture data for matching. Without any additional annotation effort, we are able to significantly outperform the current state of the art.
openaccess.thecvf.com
以上显示的是最相近的搜索结果。 查看全部搜索结果