Abstract

People with visual impairments may experience difficulties in learning new physical exercises due to a lack of visual feedback. Learning and practicing yoga is especially challenging for this population as yoga requires imitation-oriented learning. A typical yoga class requires students to observe and copy poses and movements as the instructor presents them, while maintaining postural balance during the practice. Without additional, nonvisual feedback, it can be difficult for students with visual impairments to understand whether they have accurately copied a pose – and if they have not, how to fix an inaccurate pose. Therefore, there is a need for an intelligent learning system that can capture a person’s physical posture and provide additional, nonvisual feedback to guide them into a correct pose. This study is a preliminary step toward the development of a wearable inertial sensor-based virtual learning system for people who are blind or have low vision. Using hierarchical task analysis, we developed a step-by-step conceptual model of yoga poses, which can be used in constructing an effective nonvisual feedback system. We also ranked sensor locations according to their importance by analyzing postural deviations in each pose compared to the reference starting pose.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.