TY - GEN
T1 - A unified framework for locating and recognizing human actions
AU - Xie, Yuelei
AU - Chang, Hong
AU - Li, Zhe
AU - Liang, Luhong
AU - Chen, Xilin
AU - Zhao, Debin
PY - 2011
Y1 - 2011
N2 - In this paper, we present a pose based approach for locating and recognizing human actions in videos. In our method, human poses are detected and represented based on deformable part model. To our knowledge, this is the first work on exploring the effectiveness of deformable part models in combining human detection and pose estimation into action recognition. Comparing with previous methods, ours have three main advantages. First, our method does not rely on any assumption on video preprocessing quality, such as satisfactory foreground segmentation or reliable tracking; Second, we propose a novel compact representation for human pose which works together with human detection and can well represent the spatial and temporal structures inside an action; Third, with human detection taken into consideration in our framework, our method has the ability to locate and recognize multiple actions in the same scene. Experiments on benchmark datasets and recorded cluttered videos verified the efficacy of our method.
AB - In this paper, we present a pose based approach for locating and recognizing human actions in videos. In our method, human poses are detected and represented based on deformable part model. To our knowledge, this is the first work on exploring the effectiveness of deformable part models in combining human detection and pose estimation into action recognition. Comparing with previous methods, ours have three main advantages. First, our method does not rely on any assumption on video preprocessing quality, such as satisfactory foreground segmentation or reliable tracking; Second, we propose a novel compact representation for human pose which works together with human detection and can well represent the spatial and temporal structures inside an action; Third, with human detection taken into consideration in our framework, our method has the ability to locate and recognize multiple actions in the same scene. Experiments on benchmark datasets and recorded cluttered videos verified the efficacy of our method.
UR - https://www.scopus.com/pages/publications/80052896502
U2 - 10.1109/CVPR.2011.5995648
DO - 10.1109/CVPR.2011.5995648
M3 - 会议稿件
AN - SCOPUS:80052896502
SN - 9781457703942
T3 - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
SP - 25
EP - 32
BT - 2011 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2011
PB - IEEE Computer Society
ER -