Abstract
This project addresses some research issues concerning design of intelligent user interfaces for improving human–robot interaction. In some critical applications, users interact with robots via Graphical User Interfaces (GUIs), which usually contain standard components considering a large number of users. Some of these user interface components may be redundant and sometimes confusing for some users depending on their preferences, capabilities, and the context robots are used in. This paper describes an adaptive system that enables a mobile robot to learn its users’ preferences and capabilities so that it can offer a dynamic and efficient GUI for each user rather than a standard GUI for all users. The system predicts future actions of the users by generating models based on the users’ previous interactions with the robot. The system was implemented and evaluated on a Pioneer 3-AT mobile robot. About 20 participants who were assessed on spatial ability directed the robot in simple spatial navigation tasks to evaluate effectiveness of the adaptive interface. Time to complete the task, the number of steps, and the number of errors were collected. The results showed that although spatial reasoning ability plays an important role in mobile robot navigation, it is less important in the robot control with adaptive interfaces compared to that of the non-adaptive.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.