Abstract

The present paper proposes an implementation of a hybrid hardware–software system for the visual servoing of prosthetic arms. We focus on the most critical vision analysis part of the system. The prosthetic system comprises a glass-worn eye tracker and a video camera, and the task is to recognize the object to grasp. The lightweight architecture for gaze-driven object recognition has to be implemented as a wearable device with low power consumption (less than 5.6 W). The algorithmic chain comprises gaze fixations estimation and filtering, generation of candidates, and recognition, with two backbone convolutional neural networks (CNN). The time-consuming parts of the system, such as SIFT (Scale Invariant Feature Transform) detector and the backbone CNN feature extractor, are implemented in FPGA, and a new reduction layer is introduced in the object-recognition CNN to reduce the computational burden. The proposed implementation is compatible with the real-time control of the prosthetic arm.

Highlights

  • Introduction andState-of-the ArtOne of the problems assistive robotics addresses is the production of upper limb prostheses for amputees

  • We have proposed a hybrid implementation of a visual analysis part for visual servoing of a prosthetic arm

  • The system was partitioned between the field-programmable gate arrays (FPGA) fabric and the ARM Cortex A53 processors of the Xilinx ZCU102 development board, based on the computing performance measurements of the building blocks

Read more

Summary

Introduction and State-of-the Art

One of the problems assistive robotics addresses is the production of upper limb prostheses for amputees. To overcome the limitations of traditional control solely based on the electromyographic (EMG) activity of the remaining muscles, promising alternatives consider hybrid systems combining noninvasive motion capture and vision control [1,2]. They include camera vision modules that allow for recognition of the subject’s intention to grasp an object and assist visual control of prosthetic arms for object reaching and grasping [3]. We propose a hybrid hardware/software (HW/SW) architecture for the analysis of a visual scene for the visual servoing of a neuroprosthetic arm using a glass-worn camera. The visual task here is to recognize the object the subject intends to grasp and localize it in the egocentric visual scene

State-of-the-Art lightweight CNNs for Object Detection
System Overview
Gaze Point Alignment
Noise Reduction
Gaze-Driven Object Recognition CNN
System Hybridization
Dataset
Geometric Alignment Measurements
Kernel Density Estimation
Bounding Box Generation Time Measurements
Gaze-Driven Object-Recognition CNN Time Measurements
Gaze-Driven Faster RCNN Accuracy
Time Measurement of the Whole System
Conclusions and Perspectives
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call