Back to Search
Start Over
Sequence Effects in Crowdsourced Annotations
- Source :
- EMNLP
- Publication Year :
- 2017
- Publisher :
- Association for Computational Linguistics, 2017.
-
Abstract
- Manual data annotation is a vital component of NLP research. When designing annotation tasks, properties of the annotation interface can unintentionally lead to artefacts in the resulting dataset, biasing the evaluation. In this paper, we explore sequence effects where annotations of an item are affected by the preceding items. Having assigned one label to an instance, the annotator may be less (or more) likely to assign the same label to the next. During rating tasks, seeing a low quality item may affect the score given to the next item either positively or negatively. We see clear evidence of both types of effects using auto-correlation studies over three different crowdsourced datasets. We then recommend a simple way to minimise sequence effects.
- Subjects :
- Sequence
Computer science
business.industry
Interface (Java)
media_common.quotation_subject
05 social sciences
02 engineering and technology
computer.software_genre
050105 experimental psychology
Annotation
Component (UML)
0202 electrical engineering, electronic engineering, information engineering
020201 artificial intelligence & image processing
0501 psychology and cognitive sciences
Quality (business)
Artificial intelligence
business
computer
Natural language processing
media_common
Subjects
Details
- Database :
- OpenAIRE
- Journal :
- Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing
- Accession number :
- edsair.doi...........3026fb684ea1873a5e312c235432de11