Abstract

Sentence semantic equivalence identification (SSEI) targets to measure the semantic equivalence between two sentences. To supplement limited supervision, existing methods extensively employ contrastive learning to obtain sentence semantics. Albeit considerable progress, traditional sentence-wise contrastive learning cannot grasp the diverse semantics in the polysemic sentence. To alleviate this multi-vocal issue, a Pairwise Contrastive learning method (named PairContrast) for SSEI is developed in this study to imitate the pairwise and intercompared scenarios. Specifically, two unlabelled sentences in any anchor pair are first augmented with an enhanced augmentation strategy to generate three augmented pairs. To reduce augmentation noise, a pair mix-up strategy is also employed to merge these augmented pairs into an anchor-positive pair, which is further combined with the anchor pair to pretrain the interaction module through contrastive learning. Finally, the pretrained SSEI model is finetuned on limited supervision by the binary cross entropy objective. Experiments on two publicly available SSEI datasets demonstrate the superiority of PairContrast against state-of-the-art baselines. The robustness of PairContrast under different scales of limited supervision is also verified.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call