Abstract

This paper presents a novel RGB-D 3D reconstruction algorithm for the indoor environment. The method can produce globally-consistent 3D maps for potential GIS applications. As the consumer RGB-D camera provides a noisy depth image, the proposed algorithm decouples the rotation and translation for a more robust camera pose estimation, which makes full use of the information, but also prevents inaccuracies caused by noisy depth measurements. The uncertainty in the image depth is not only related to the camera device, but also the environment; hence, a novel uncertainty model for depth measurements was developed using Gaussian mixture applied to multi-windows. The plane features in the indoor environment contain valuable information about the global structure, which can guide the convergence of camera pose solutions, and plane and feature point constraints are incorporated in the proposed optimization framework. The proposed method was validated using publicly-available RGB-D benchmarks and obtained good quality trajectory and 3D models, which are difficult for traditional 3D reconstruction algorithms.

Highlights

  • Complete 3D models of the indoor environment are an important part of modern digital city systems [1,2]

  • We propose an algorithm that is robust enough for a flexible hand-held RGB-D camera scanning the indoor environment

  • This paper presents a novel 3D reconstruction algorithm for the indoor environment using consumer RGB-D cameras

Read more

Summary

Introduction

Complete 3D models of the indoor environment are an important part of modern digital city systems [1,2]. As every frame observes hundreds of visual features and the frames increase linearly with the size of the environment or the scanning time, the algorithm cannot handle all the variables in one optimization problem One solution to this issue is to use a keyframe scheme. The algorithm only utilizes the feature points from RGB images and produces an accurate rotation estimation and a rotation without scale It calculates a one-to-one correspondence for all the pixels in the matching frames and estimate an absolute translation using all the depth information.

Related Work
Visual Feature Detection
Uncertainty of Depth Measurements
Rotation Solving
Absolute Translation Recovery
Back Projection Associations
Solve Translation
Plane Constraints
Plane Extraction
Plane Points’ Associations
Joint Optimization
Experimental Section
Findings
Conclusions and Future Work
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.