Abstract
Meta-learning-based methods prevail in few-shot text classification. Current methods perform meta-training and meta-testing on two parts of a dataset in the same or similar domains. This results in a significant limit in model performance when faced with data from different domains, limiting the generalization of few-shot models. To solve this problem, this study proposes a new setting, namely, domain-generalized few-shot text classification. First, meta-training is conducted on a multi-domain dataset to learn a generalizable model. Subsequently, the model is meta-tested on a target dataset. In addition, a domain-generalized model, namely, a dual adversarial network, is designed to improve the meta-learning-based methods under domain drift between different datasets and domains. Unlike previous meta-learning methods, two N-way-K-shot tasks were input from different domains for a dual adversarial network at each episode. Dual adversarial networks leverage the features from two different domains for adversarial training to improve the domain adaptability of the model. The proposed model utilizes a domain-knowledge generator during adversarial training to produce domain-specific knowledge, and a domain discriminator to recognize the domain label of the produced knowledge. Extensive experiments are conducted to verify the effectiveness of the proposed settings and model. The experimental results show that the model performance in our proposed setting is improved by an average of 3.84% compared to that in cross-domain few-shot text classification. Furthermore, the dual adversarial network significantly outperforms the five competitive baseline models, with an average improvement of 7.20%. The proposed model achieves an average performance improvement of 2.69% compared with the best baseline method.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.