Abstract

3D pose estimation remains a challenging task since human poses exhibit high ambiguity and multi-granularity. Traditional graph convolution networks (GCNs) accomplish the task by modeling all skeletons as an entire graph, and are unable to fuse combinable part-based features. By observing that human movements occur due to part of human body (i.e. related skeletons and body components, known as the poselet) and those poselets contribute to each movement in a hierarchical fashion, we propose a hierarchical poselet-guided graph convolutional network (HPGCN) for 3D pose estimation from 2D poses. HPGCN sets five primitives of human body as basic poselets, and constitutes high-level poselets according to the kinematic configuration of human body. Moreover, HPGCN forms a fundamental unit by using a diagonally dominant graph convolution layer and a non-local layer, which corporately capture the multi-granular feature of human poses from local to global perspective. Finally HPGCN designs a geometric constraint loss function with constraints on lengths and directions of bone vectors, which help produce reasonable pose regression. We verify the effectiveness of HPGCN on three public 3D human pose benchmarks. Experimental results show that HPGCN outperforms several state-of-the-art methods.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.