Abstract

Document image classification is a challenging task due to the complexity of information contained within documents, including text, images, and their spatial arrangement. Deep learning has become a pivotal tool for extracting and learning complex patterns. However, conventional methods often grapple with integrating different data modalities and minimizing redundancy, leading to a need for more advanced and efficient deep learning strategies. This study presents a new approach to document image classification, named graph attention-driven with dual tune learning (GAD-DTL). GAD-DTL employs dual-tune learning and graph attention networks. The methodology creates semantic region embedding within document images, which incorporate both textual and spatial data. A key feature of this approach is the adaptive fusion layer, which integrates different modalities and uses a graph attention layer to capture context within each region. To minimize redundancy in learned features, we implement two distinct learning techniques, relational and non-relational learning. This approach enhances document image classification by ensuring invariant representation and minimal redundancy in features.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call