Abstract
Vision navigation based on scene matching between real-time images and a reference image has many advantages over the commonly used inertial navigation system (INS), such as no cumulative measurement errors for long-endurance flight. Recent developments in vision navigation are mainly used for partial navigation parameters measurements, such as the position and the relative velocity, which cannot meet the requirements of completely autonomous navigation. We present the concept, principle and method of full-parameter vision navigation (FPVN) based on scene matching. The proposed method can obtain the three-dimensional (3D) position and attitude angles of an aircraft by the scene matching for multiple feature points instead of a single point in existing vision navigations. Thus, FPVN can achieve the geodetic position coordinates and attitude angles of the aircraft and then the velocity vector, attitude angular velocity and acceleration can be derived by the time differentials, which provide a full set of navigation parameters for aircrafts and unmanned aerial vehicles (UAVs). The method can be combined with the INS and the cumulative errors of the INS can be corrected using the measurements of FPVN rather than satellite navigation systems. The approach provides a completely autonomous and accurate navigation method for long-endurance flight without the help of satellites.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.