People with severe speech and motor impairment often find it difficult to manipulate physical objects due to spasticity and have familiarity with eye pointing based communication. This article presents a novel eye gaze controlled augmented reality human-robot interface that maintains a safe distance of the robot from the operator. We used a bespoke appearance-based eye gaze tracking algorithm and compared two different safe distance maintenance algorithms. We undertook simulation studies followed by a user trial involving end users. Users with severe speech and motor impairment could bring the robotic arm at any designated point within its working envelope in less than 3 minutes.
Severe Speech Severe Speech Impairment Severe Motor Impairment Motor Impairment Safe Maintenance Speech Impairment Severe Motor Safe Distance Eye Gaze Appearance-based Gaze
AI-powered Research feed
Introducing Weekly Round-ups!Beta
Round-ups are the summaries of handpicked papers around trending topics published every week. These would enable you to scan through a collection of papers and decide if the paper is relevant to you before actually investing time into reading it.
Climate change Research Articles published between Sep 12, 2022 to Sep 18, 2022
Sep 19, 2022
Articles Included: 5
Rainfall projections from the Coupled Model Intercomparison Project (CMIP) models are strongly tied to projected sea surface temperature (SST) spatial...Read More
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on “as is” basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The Copyright Law.