Abstract

Standard test sets for supervised learning evaluate in-distribution generalization. Unfortunately, when a dataset has systematic gaps (e.g., annotation artifacts), these evaluations are misleading: a model can learn simple decision rules that perform well on the test set but do not capture a dataset's intended capabilities. We propose a new annotation paradigm for NLP that helps to close systematic gaps in the test data. In particular, after a dataset is constructed, we recommend that the dataset authors manually perturb the test instances in small but meaningful ways that (typically) change the gold label, creating contrast sets. Contrast sets provide a local view of a model's decision boundary, which can be used to more accurately evaluate a model's true linguistic capabilities. We demonstrate the efficacy of contrast sets by creating them for 10 diverse NLP datasets (e.g., DROP reading comprehension, UD parsing, IMDb sentiment analysis). Although our contrast sets are not explicitly adversarial, model performance is significantly lower on them than on the original test sets---up to 25\% in some cases. We release our contrast sets as new evaluation benchmarks and encourage future dataset construction efforts to follow similar annotation processes.

Highlights

  • Progress in natural language processing (NLP) has long been measured with standard benchmark datasets (e.g., Marcus et al, 1993)

  • Findings of the Association for Computational Linguistics: EMNLP 2020, pages 1307–1323 November 16 - 20, 2020. c 2020 Association for Computational Linguistics fills in these systematic gaps in the test set

  • We propose that dataset authors manually perturb instances from their test set, creating contrast sets which characterize the correct decision boundary near the test instances (Section 2)

Read more

Summary

Introduction

Progress in natural language processing (NLP) has long been measured with standard benchmark datasets (e.g., Marcus et al, 1993). These benchmarks help to provide a uniform evaluation of new modeling developments. Example Textual Perturbations: Two -colored and -posed cats are face to face in one image. Three -colored and -posed chow dogs are face to face in one image. Two differently-colored but -posed chow dogs are face to face in one image. Example Image Perturbation: Two -colored and -posed chow dogs are face to face in one image

Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.