作者
Weiyu Zhang, Menglong Zhu, Konstantinos G Derpanis
发表日期
2013
研讨会论文
International Conference on Computer Vision
简介
This paper presents a novel approach for analyzing human actions in non-scripted, unconstrained video settings based on volumetric, xyt, patch classifiers, termed actemes. Unlike previous action-related work, the discovery of patch classifiers is posed as a strongly-supervised process. Specifically, keypoint labels (eg, position) across spacetime are used in a data-driven training process to discover patches that are highly clustered in the spacetime keypoint configuration space. To support this process, a new human action dataset consisting of challenging consumer videos is introduced, where notably the action label, the 2D position of a set of keypoints and their visibilities are provided for each video frame. On a novel input video, each acteme is used in a sliding volume scheme to yield a set of sparse, non-overlapping detections. These detecsseddeetecctions provide the intermediate substrate for segmeegmenatot the action. For action classification, the proposed representation shows significant improvement over state-of-the-art low-level features, while providing spatiotemporal localization as additional output. This output sheds further light into detailed action understanding.
引用总数
20152016201720182019202020212022202320249141740475762687035
学术搜索中的文章