With the advancement of 3D scanning technologies and deep learning theories, point cloud-based deep learning networks have gained considerable attention in the fields of 3D vision and computer graphics. Leveraging the rich geometric information present in 3D point clouds, these networks facilitate more accurate feature learning tasks. However, existing networks often suffer from generalization defects caused by variations in pose and inconsistent representations of training data. In this paper, we propose a novel data augmentation framework to overcome these limitations. Our approach utilizes principal component analysis (PCA) to generate four aligned copies of a point cloud. These copies are then input into a multi-channel structure, which is compatible with popular backbones of point cloud-based deep networks. Finally, the outputs of the multi-channel structure are merged to generate rotation-invariant feature learning results. Experimental evaluations demonstrate the efficacy of our framework, showcasing significant improvements in various existing point cloud-based deep learning methods. Notably, our method exhibits enhanced robustness in classification tasks, particularly when dealing with point clouds containing random pose variations and non-uniform densities. Project link: https://github.com/LAB123-tech/PCAlign.
Read full abstract