Abstract

Domain gaps between synthetic text and real-world text restrict current text recognition methods. One solution is to align features through Unsupervised Domain Adaptation (UDA). Most existing UDA-based text recognition methods extract global and local features to alleviate domain differences, only focusing on character-level distribution gaps. However, notable distribution gaps in character combinations exert a pivotal influence on diverse text recognition tasks. To this end, we propose a Multi-level And multi-Granularity domain adaptation with entropy loss guIded text reCognition model, named MAGIC. It integrates Global-level Domain Adaptation (GDA) to mitigate image-level domain drift and Local-level Multi-granularity Domain Adaptation (LMDA) for local feature shifts. Particularly, we design a subword-level domain discriminator to align the subword features relating to each character combination. Moreover, multi-granularity entropy minimization is used to optimize the target domain data for better domain adaptation. Experimental results on several types of text datasets demonstrate the effectiveness of MAGIC.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.