Abstract

Vision navigation based on scene matching between real-time images and a reference image has many advantages over the commonly used inertial navigation system (INS), such as no cumulative measurement errors for long-endurance flight. Recent developments in vision navigation are mainly used for partial navigation parameters measurements, such as the position and the relative velocity, which cannot meet the requirements of completely autonomous navigation. We present the concept, principle and method of full-parameter vision navigation (FPVN) based on scene matching. The proposed method can obtain the three-dimensional (3D) position and attitude angles of an aircraft by the scene matching for multiple feature points instead of a single point in existing vision navigations. Thus, FPVN can achieve the geodetic position coordinates and attitude angles of the aircraft and then the velocity vector, attitude angular velocity and acceleration can be derived by the time differentials, which provide a full set of navigation parameters for aircrafts and unmanned aerial vehicles (UAVs). The method can be combined with the INS and the cumulative errors of the INS can be corrected using the measurements of FPVN rather than satellite navigation systems. The approach provides a completely autonomous and accurate navigation method for long-endurance flight without the help of satellites.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call