Abstract

In competitive programming, standard solutions for easy tasks are usually simple and shorter, making submissions more convergent both in idea and texts. The huge difference in submission diversity between easy and hard tasks, brings inescapable challenges to plagiarism judging by means of similarity thresholding. In this paper, by drawing the strong data support from the China National Olympiads in Informatics (NOI), we study the statistical characteristics of submission similarities for tasks of wide range of difficulty degrees. Finally, we propose a new adaptive method to detect submission plagiarism as out-of-distribution samples, together with a large-scale challenge dataset of competitive programming submission plagiarism detection. Our method is shown to be of higher accuracy and robustness, thus feasible and reliable for largescale competitive programming contests.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call