Abstract
The challenge in addressing uncalibrated visual servoing (VS) control of robot manipulators with unstructured environments is to obtain appropriate interaction matrix and keep the image features in the field of view (FOV), especially when the non-Gaussian noise disturbance exists in the VS process. In this article, a hybrid control algorithm which combines bidirectional extreme learning machine (B-ELM) with smooth variable structure filter (SVSF) is proposed to estimate interaction matrix and tackle visibility constraints. For VS, the nonlinear mapping between image features and interaction matrix is approximated using the B-ELM learning. To increase the capability of anti-interference, the SVSF is employed to re-estimate interaction matrix. A constraint function presenting feature coordinates and region boundaries is given and added to the velocity controller, which drags image features away from the restricted region and ensures the smoothness of the velocities. Since the camera and robot model parameters are not required in developing the control strategy, the servoing task can be fulfilled flexibly and simply. Simulation and experimental results on a conventional 6-degree-of-freedom manipulator verify the effectiveness of the proposed method.
Highlights
Visual feedback signals have been used as significant information in robots to tackle the positioning or motion control in unstructured environments
According to the visual feedback signal returned by 3D Cartesian space coordinate or image plane coordinate, it can be divided into position-based visual servoing (PBVS), image-based visual servoing (IBVS) and hybrid visual servoing control systems [1]–[3]
A method combined bidirectional extreme learning machine (B-ELM) algorithm with smooth variable structure filter (SVSF) algorithm was proposed to map the nonlinearity between the visual space of the robot manipulator and the motion space of the end-effector, and the field of view (FOV) constraint was adopted to ensure that no feature points will be lost during the whole motion
Summary
Visual feedback signals have been used as significant information in robots to tackle the positioning or motion control in unstructured environments. In IBVS, the analytical form of the image Jacobian matrix is obtained in the control loop which requires accurate calibration of camera model parameters. An uncalibrated IBVS, which does not require scene model or camera/robot calibration, applies image feature information (e.g., point feature, line feature and image moment) to estimate the unknown system dynamics, and the controller is designed based on the identified Jacobian matrix [4], [5]. A method combined bidirectional extreme learning machine (B-ELM) algorithm with smooth variable structure filter (SVSF) algorithm was proposed to map the nonlinearity between the visual space of the robot manipulator and the motion space of the end-effector, and the FOV constraint was adopted to ensure that no feature points will be lost during the whole motion. Assumption 2: Image feature points observed by camera are coplanar
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.