Abstract

Intelligent systems, like driver assistance systems, remain within vehicles and help drivers by providing essential information about traffic, blockage of roads, and possible routes for safe driving. The objective of scene text spotting in a driver assistance system is to localize and recognize scene texts, signs of milestones, traffic panels, and road marks in natural scene images. However, text edges get fainted due to adverse weather conditions, like fog, rain, smog, or poor contrast. This makes the task of spotting more challenging. In this paper, we propose an end-to-end trainable deep neural network, known as <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">Met-MLTS</i> , that can address the issue of spotting multi-oriented text instances in scene images captured in adverse meteorological conditions. It localizes words, predicts script class, and performs word spotting for every rotated bounding box. It is a fast multilingual scene text spotter that utilizes hierarchical spatial context, channel-wise inter-dependencies, and semantic edge supervision to localize and recognize words and predict script class in scene images using smartphones. We explore inter-class interference to reduce the misclassification problem. A light-weight recognition module for multilingual character segmentation, word-level recognition, and script identification is incorporated. We demonstrate the efficacy of our spotting network on resource-constraint devices.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call