Abstract
Zero-shot stance detection (ZSSD) aims to classify stances on unseen topic data with the ability of a model to generalize in the range of topics in the real world. The key point of ZSSD is to prevent the model from overfitting on seen topic data and show robust stance recognition performance on unseen topic samples. In this paper, we propose a new ZSSD framework to generalize stance detection performance by alleviating seen data information from the encoder and focusing on improving stance classification ability in zero-shot conditions. The proposed framework involves two learning stages contributing to ZSSD: topic-agnostic text encoder learning and zero-shot meta-learning. Our framework achieves notable improvements on the three benchmark zero-shot stance detection datasets and zero-shot aspect target sentiment classification dataset showing the effectiveness of our method in the zero-shot settings.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.