Abstract

We propose a novel deep learning-based inertial odometry framework to improve both localization accuracy and efficiency. The measured 3D accelerations and angular velocities are constructed as two independent channel 2D features. In each channel, the 3 axes of accelerations or angular velocities constitute the height dimension, whereas the temporal axis represents the width dimension. In the neural network module, we first employ 2D convolutional operations to convert the input features to 1D series with 4-fold downsampling along the temporal axis. We then apply the convolutional attention module to refine the generated 1D feature map and long short-term memory (LSTM) layers to generate the hidden state for each refined time index, which is followed by two separate 2-layer multilayer perceptrons (MLPs) to regress the velocity and the uncertainty. During backpropagation, we use the negative log-likelihood loss to optimize the velocity and the uncertainty objectives. Benefiting from the velocity sequence objective in a single network forward propagation procedure, we also add the average squared distance between the ground-truth relative position and that of the prediction to minimize the location error. Extensive experiments on both the RoNIN dataset and the real scenario collected data in the CUHK campus show higher localization accuracy with 10 times higher efficiency of the proposed framework than the comparing methods.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call