Abstract

The detection of abnormal postures, such as that of a reclining person, is a crucial part of visual surveillance. Further, even regular poses can appear rotated because of incongruity between the image and the angle of a pre-installed camera. However, most existing human pose estimation methods focus on small rotational changes, i.e., those less than 50 degrees, and they seldom consider robust human pose estimation for more drastic rotational changes. To the best of our knowledge, there have been no reports on the robustness of human pose estimation for rotational changes through large angles. In this study, we propose a robust human pose estimation method by creating a path for learning new rotational changes based on a self-supervised method and by combining the results with those obtained from a path based on a supervised method. Furthermore, a combination module composed of a convolutional layer is trained complementarily by both paths of the network to produce robust results for various rotational changes. We demonstrate the robustness of the proposed method with extensive experiments on images generated by rotating the elements of standard benchmark datasets. We fully analyze the rotational characteristics of the state-of-the-art human pose estimators and the proposed method. On the COCO Keypoint Detection dataset, the proposed method attains more than 15% improvement in the mean of average precision compared to the state-of-the-art method, and the standard deviation of the performance is improved by more than 4.7 times.

Highlights

  • The problem of estimating a human pose based on a single image can be reduced to accurately locating a set of semantic keypoints, e.g., the head, shoulder, elbow, wrist, knee, and ankle of a human body [1]

  • We propose a robust human pose estimation method by creating a new path for learning rotational changes based on a self-supervised method and combining it with the results of a path based on a supervised method

  • In this study, we analyzed the robustness of human pose estimation with respect to rotational changes through large angles and proposed a novel method to improve rotational robustness

Read more

Summary

Introduction

The problem of estimating a human pose based on a single image can be reduced to accurately locating a set of semantic keypoints, e.g., the head, shoulder, elbow, wrist, knee, and ankle of a human body [1]. This is a fundamental task in more complicated problems such as action recognition [2], [3], and it is useful in various applications such as human-computer interaction and animation [4], [5]. We focus on single person pose estimation because it is the basic step in other applications, such as multi-person pose estimation, video-based human pose estimation, and tracking in two-dimensional or three-dimensional spaces.

Objectives
Methods
Findings
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call