Abstract
This paper presents a robust and fully-automatic human motion tracking system without motion priors information using a camera in a fixed location. Bottom-up estimation approaches have recently been applied to such tasks with some success. However, the performance of these approaches is limited by the difficulty of building an effective appearance model. In particular, the appearance model must be derived from initial estimates of the tracked person's limb posture. However, in addition to inaccuracies in this initial estimate, the precise shape, size and boundaries of the tracked person's limbs are not known. Hence it is inevitable that background (non-limb) pixels are included into the appearance model. In the case of smaller limbs such as the arms, this can cause the model to become unrepresentative and sometimes confused with other body parts such as the torso. In this paper, we address the problem of how to automatically extract accurate training samples for building an accurate appearance model, and propose a mechanism for identifying and removing background (negative) pixels via pixel clustering that is robust even with a loose-fitting body shape model. Experiments are conducted to compare the proposed approach against existing appearance-based algorithms without negative pixel removal using several publicly available data sets. Results show that tracking accuracy is consistently improved, and significantly so for small limbs such as the arms.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.