Improving Human-Labeled Data through Dynamic Automatic Conflict Resolution
@inproceedings{Sun2020ImprovingHD, title={Improving Human-Labeled Data through Dynamic Automatic Conflict Resolution}, author={David Q. Sun and Hadas Kotek and C. Klein and Mayank Gupta and William Li and J. Williams}, booktitle={COLING}, year={2020} }
This paper develops and implements a scalable methodology for (a) estimating the noisiness of labels produced by a typical crowdsourcing semantic annotation task, and (b) reducing the resulting error of the labeling process by as much as 20-30% in comparison to other common labeling strategies. Importantly, this new approach to the labeling process, which we name Dynamic Automatic Conflict Resolution (DACR), does not require a ground truth dataset and is instead based on inter-project… CONTINUE READING
References
SHOWING 1-10 OF 42 REFERENCES
Cheap and Fast - But is it Good? Evaluating Non-Expert Annotations for Natural Language Tasks
- Computer Science
- EMNLP
- 2008
- 1,989
- PDF
Repeated labeling using multiple noisy labelers
- Computer Science
- Data Mining and Knowledge Discovery
- 2013
- 144
- PDF
The Benefits of a Model of Annotation
- Computer Science
- Transactions of the Association for Computational Linguistics
- 2014
- 123
- PDF