The Next-Best-View (NBV) algorithm is a key component in autonomous unknown object reconstruction. It iteratively determines the optimal sensor pose to capture the maximum information about the object under reconstruction. However, prevailing deep reinforcement learning (DRL) based NBV algorithms tend to transform point cloud, the raw sensor data, into different representations, thereby obscuring natural invariances of the data. In this work, we propose an innovative DRL-based method, denoted as RL-NBV, to learn NBV policy directly from the raw point cloud data. Specifically, we interpret the observation space as the current state of the reconstructed object represented by point clouds and current view selection states. Experimental results indicate that our method outperforms existing methods in terms of reconstruction performance. Moreover, our method significantly improves efficiency over ray-casting-based algorithms as time-consuming ray casting and data transformation are unnecessary.
Read full abstract