Abstract

The need for the annotated training dataset on which data-hungry machine learning algorithms feed has increased dramatically with advanced acclaim of machine learning applications. To annotate the data, people with domain expertise are needed, but they are seldom available and expensive to hire. This has lead to the thriving of crowdsourcing platforms such as Amazon Mechanical Turk (AMT). However, the annotations provided by one worker cannot be used directly to train the model due to the lack of expertise. Existing literature in annotation aggregation focuses on binary and multi-choice problems. In contrast, little work has been done on complex tasks such as sequence labeling with imbalanced classes, a ubiquitous task in Natural Language Processing (NLP), and Bio-Informatics. We propose OptSLA, an Optimization-based Sequential Label Aggregation method, that jointly considers the characteristics of sequential labeling tasks, workers reliabilities, and advanced deep learning techniques to conquer the challenge. We evaluate our model on crowdsourced data for named entity recognition task. Our results show that the proposed OptSLA outperforms the state-of-the-art aggregation methods, and the results are easier to interpret.

Highlights

  • Crowdsourcing (Howe, 2008) is a popular platform to annotate massive corpora inexpensively

  • We study the annotation aggregation problem for sequential labeling tasks, a common Natural Language Processing (NLP) task

  • To evaluate the proposed optimization-based sequential label aggregation method (OPTSLA), we compare the span level precision, recall, and F1 score3 of the inferred aggregation labels with three state-of-theart baselines methods hidden Markov model (HMM)-crowd (Nguyen et al, 2017), Conditional Random Fields (CRF)-MA (Rodrigues et al, 2014), and BSCseq result comes from (Simpson and Gurevych, 2019)

Read more

Summary

Introduction

Crowdsourcing (Howe, 2008) is a popular platform to annotate massive corpora inexpensively. It has bred lots of interest in machine learning and deep learning tasks. One common annotation aggregation approach is Majority Voting (MV) (Lam and Suen, 1997), in which annotation with the highest number of occurrences is deemed as truth. Another naive approach is to regard an annotation as correct if a certain number of workers provide the same annotation. We study the annotation aggregation problem for sequential labeling tasks, a common NLP task

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call