Abstract

The textual data of a document is supplemented by the graphical information in it. To make communication easier, they contain tables, charts and images. However, it excludes a section of our population - the visually impaired. With technological advancements, the blind can access the documents through text to speech software solutions. In this method, even images can be conveyed by reading out the figure captions. However, charts and other statistical comparisons which involve critical information are difficult to be “read” out this way. Aim of this paper is to analyse various methods available to solve this vexatious issue. We survey the state-of-the-art works that do the exact opposite of graphing tools. In this paper, we explore the existing literature in understanding the graphs and extracting the visual encoding from them. We classify these approaches into modality-based approaches, conventional and deep-learning based methods. The survey also contains comparisons and analyses relevant study datasets. As an outcome of this survey, we observe that: (i) All existing works under each category need decoding in a variety of graphs. (ii) Among the approaches, deep learning performs remarkably well in localisation and classification. However, it needs further improvements in reasoning from chart images. (iii) Research works are still in progress to access data from vector images. Recreating data from the raster images has unresolved issues. Based on this study, the various applications of decoding the graphs, challenges and future possibilities are also discussed. This paper explores current works in the extraction of chart data, which seek to enable researchers in Human Computer Interaction to achieve human-level perception of visual data by machines. In this era of visual summarisation of data, the AI approaches can automate the underlying data extraction and hence provide the natural language descriptions to support visually disabled users.

Highlights

  • The recent advancements in Assistive Technology (AT) has revolutionised the way cognitively limited users interact with the world

  • Some works concentrate on recognizing the figure in a document as a graphical image [45] or not, while others classify it into various types of charts [4], [28], [29], [39], whereas others concentrate on extracting the visual elements from the charts [26], [40], [46]

  • The metrics for evaluating object proposals are based on intersection over union (IOU) given

Read more

Summary

Introduction

The recent advancements in Assistive Technology (AT) has revolutionised the way cognitively limited users interact with the world. Visual impairment stands out as the most limiting amongst these disabilities. Assistive technology is defined as any technology that is built to help a person with disability. The advances in Artificial Intelligence is tremendous, paving the way to autonomous vehicles. We can make efficient use of these algorithms in Assistive Technology as well, to help the visually impaired ones to enhance in the field of education, navigation and to improve social interac-. Disabled persons have access to information by touch or voice. Connier et al [1] discusses the incorporation of different assistive technology devices into the Internet of Things (IoT). A new model is introduced, linking an AT device with Smart Objects and their cloud to enhance VIP’s perception

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call