Abstract

Unstructured point clouds are a representative shape representation of real-world scenes in 3D vision and graphics. Incompletion inevitably arises, due to the way the set of unorganized points is captured, e.g., as fusion of depth images, merged laser scans, or structure-from-x. In this paper, an end-to-end sparse-to-dense multi-encoder neural network (termed an SDME-Net) is proposed for uniformly completing an unstructured point cloud with its shape details preserved. Unlike most existing learning-based shape completion methods that are enforced on the representations of 2D images and 3D voxelization of point clouds, and require priors of the underlying shape's structures, topologies and annotations, the SDME-Net is implemented on the incomplete and even noisy point cloud without any transformation, and makes no specific assumptions about the incompletion distribution and geometry features in the input. Specifically, the defective point cloud is completed and optimized in a sparse-to-dense manner of two-stages. In the first stage, we generate a sparse but complete point cloud based on a bistratal PointNet, and in the second stage, we yield a dense and high-fidelity point cloud by encoding and decoding the sparse result in the first stage using PointNet++. Meanwhile, we combine the distance loss and repulsion loss to generate more uniformly distributed output point clouds closer to the ground-truth counterparts. Qualitative and quantitative experiments on the public ShapeNet dataset illustrate that our approach outperforms the state-of-art learning-based point cloud shape completion methods in terms of real structure recovery, uniformity, and noise/partiality robustness.

Highlights

  • Since early 1985s, point cloud has been recognized as a representative form of 3D objects which is widely used as the standard output of various sensors [1]

  • Smart geometry processing of point clouds again entered into the spot of 3D vision and graphics, due to the rapid advances and applications of artificial intelligence (AI) in robotics, autonomous driving and mixed reality [2]–[4]

  • NETWORK ARCHITECTURE As shown in Fig. 2, the whole network architecture of the SDME-Net is two continuous point cloud coders, where N and Ni are the number of points, Ci are the number of feature channels, and the colored rounded rectangles represent certain baseline network architectures

Read more

Summary

INTRODUCTION

Since early 1985s, point cloud has been recognized as a representative form of 3D objects which is widely used as the standard output of various sensors [1]. Neural network methods for processing point cloud data are increasingly used in 3D target recognition, scene reconstruction [13]–[16], target object completion [6], [7], [17], [18], and 3D shape representation learning [13], [16], [19]. A recent work [31] proposed an end-to-end point cloud completion network, and the network directly processes the incomplete input point cloud to generate complete point clouds, showing excellent performance and far superior to the methods based on volume representation in terms of cost and accuracy of 3D shape generation.

STAGE I
LOSS FUNCTION
EXPERIMENT
Findings
CONCLUSION

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.