Combining local appearance and holistic view: Dual-Source Deep Neural Networks for human pose estimation

Abstract

We propose a new learning-based method for estimating 2D human pose from a single image, using Dual-Source Deep Convolutional Neural Networks (DS-CNN). Recently, many methods have been developed to estimate human pose by using pose priors that are estimated from physiologically inspired graphical models or learned from a holistic perspective. In this paper, we propose to integrate both the local (body) part appearance and the holistic view of each local part for more accurate human pose estimation. Specifically, the proposed DS-CNN takes a set of image patches (category-independent object proposals for training and multi-scale sliding windows for testing) as the input and then learns the appearance of each local part by considering their holistic views in the full body. Using DS-CNN, we achieve both joint detection, which determines whether an image patch contains a body joint, and joint localization, which finds the exact location of the joint in the image patch. Finally, we develop an algorithm to combine these joint detection/localization results from all the image patches for estimating the human pose. The experimental results show the effectiveness of the proposed method by comparing to the state-of-the-art human-pose estimation methods based on pose priors that are estimated from physiologically inspired graphical models or learned from a holistic perspective.

DOI: 10.1109/CVPR.2015.7298740

Extracted Key Phrases

0204060201520162017
Citations per Year

73 Citations

Semantic Scholar estimates that this publication has 73 citations based on the available data.

See our FAQ for additional information.

Cite this paper

@article{Fan2015CombiningLA, title={Combining local appearance and holistic view: Dual-Source Deep Neural Networks for human pose estimation}, author={Xiaochuan Fan and Kang Zheng and Yuewei Lin and Song Wang}, journal={2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, year={2015}, pages={1347-1355} }