Abstract

Much research has been done worldwide on support systems for visually impaired persons. There are still many problems in representing the real-time information that changes around the user. In this paper, we outline the design of a visual support system that provides three-dimensional visual information using three-dimensional virtual sound. Three-dimensional information is obtained by analyzing images captured by stereo cameras, and recognizing the objects needed by the blind user. Using the three-dimensional virtual acoustic display, which relies on Head Related Transfer Functions (HRTFs), the user is informed of the locations and movements of objects. The user's auditory sense is not impeded as we use a bone conduction headset, which does not block out environmental sound. The proposed system is expected to be useful in the situation where the infrastructure is incomplete, and when the situation changes in real-time. We plan to experiment with it, for example, to guide users in walking and playing sports. In acoustic interface experiments, we found that there were many mis-recognitions between front and back directions, and that active operation is needed to recognize the sound position accurately.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call