Abstract

State estimation is crucial for enabling autonomous mobility in mobile robots. However, traditional localization methods often falter in degraded environments, including issues like visual occlusion, lidar performance degradation, and global navigation satellite system signal interference. This paper presents a novel estimation approach for wheeled robots, exclusively utilizing proprioceptive sensors such as encoders and inertial measurement units (IMU). Initially, the motion manifolds extracted from the historical trajectories are used to assist the encoder in realizing the orientation estimation. Furthermore, a hybrid neural network is designed to categorize the robot’s operational state, and the corresponding pseudo-constraints are added to improve the estimation accuracy. We utilize an error state Kalman filter for the encoder and IMU data fusion. Lastly, comprehensive testing is conducted using both datasets and real-world robotic platforms. The findings underscore that the integration of manifold and motion constraints within our proposed state estimator substantially elevates accuracy compared to conventional approaches. Compare with the methods commonly used in engineering, the accuracy of this method is improved by more than 20%. Crucially, this methodology enables dependable estimation even in degraded environments.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call