Abstract

To autonomously move and operate objects in cluttered indoor environments, a service robot requires the ability of 3D scene perception. Though 3D object detection can provide an object-level environmental description to fill this gap, a robot always encounters incomplete object observation, recurring detections of the same object, error in detection, or intersection between objects when conducting detection continuously in a cluttered room. To solve these problems, we propose a two-stage 3D object detection algorithm which is to fuse multiple views of 3D object point clouds in the first stage and to eliminate unreasonable and intersection detections in the second stage. For each view, the robot performs a 2D object semantic segmentation and obtains 3D object point clouds. Then, an unsupervised segmentation method called Locally Convex Connected Patches (LCCP) is utilized to segment the object accurately from the background. Subsequently, the Manhattan Frame estimation is implemented to calculate the main orientation of the object and subsequently, the 3D object bounding box can be obtained. To deal with the detected objects in multiple views, we construct an object database and propose an object fusion criterion to maintain it automatically. Thus, the same object observed in multi-view is fused together and a more accurate bounding box can be calculated. Finally, we propose an object filtering approach based on prior knowledge to remove incorrect and intersecting objects in the object dataset. Experiments are carried out on both SceneNN dataset and a real indoor environment to verify the stability and accuracy of 3D semantic segmentation and bounding box detection of the object with multi-view fusion.

Highlights

  • In an indoor environment, objects are regarded as the main contents and they provide crucial clues for scene understanding and environmental perception

  • We propose a two-stage 3D object detection framework by fusing multiple views of a 3D point cloud based on a real-time visual SLAM for an indoor service robot

  • We propose an object filtering approach based on prior knowledge including size and volume ratio to remove atypical and intersecting objects in the object dataset

Read more

Summary

Introduction

Objects are regarded as the main contents and they provide crucial clues for scene understanding and environmental perception. Object detection helps the indoor service robot possess higher semantic awareness of its operating environment. It is difficult to detect these methods are still not enough for a robot to operate in 3D space. Notrobust robustenough enoughtotobe beused usedby byaa robot robot to to perform tasks tasks such such as as obstacle obstacle avoidance avoidance navigation navigation or or essential essential object object grabbing To resolve resolve this this problem, problem, 3D object object detection detection emerges emerges as as aa candidate candidate to to realize realize object object classification classification and and detection detection with with the the position position and information. Most methods utilize single RGB-D image of the same object in the NYU. We explore the multi-view fusion method to resolve the incomplete point the incomplete point cloud information of objects

Object
Related Work
C Pn should be transformed
Unsupervised Segmentation of the Object Point Cloud
The First Time to Insert Objects to the Object Database
Object Fusion Criterion and Database Maintenance
Object Database Refinement
Atypical Object Filtering Based on Prior Knowledge
Intersection Object Filtering Based on Volume Ratio
Different
Experimental Evaluation
Object-Level 3D Semantic Segmentation Evaluation
Evaluation
Conclusions

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.