Back to Search Start Over

Cross-language transfer of semantic annotation via targeted crowdsourcing: task design and evaluation.

Authors :
Stepanov, Evgeny A.
Chowdhury, Shammur Absar
Bayer, Ali Orkan
Ghosh, Arindam
Klasinas, Ioannis
Calvo, Marcos
Sanchis, Emilio
Riccardi, Giuseppe
Source :
Language Resources & Evaluation. Mar2018, Vol. 52 Issue 1, p341-364. 24p.
Publication Year :
2018

Abstract

Modern data-driven spoken language systems (SLS) require manual semantic annotation for training spoken language understanding parsers. Multilingual porting of SLS demands significant manual effort and language resources, as this manual annotation has to be replicated. Crowdsourcing is an accessible and cost-effective alternative to traditional methods of collecting and annotating data. The application of crowdsourcing to simple tasks has been well investigated. However, complex tasks, like cross-language semantic annotation transfer, may generate low judgment agreement and/or poor performance. The most serious issue in cross-language porting is the absence of reference annotations in the target language; thus, crowd quality control and the evaluation of the collected annotations is difficult. In this paper we investigate <italic>targeted</italic> crowdsourcing for semantic annotation transfer that delegates to crowds a complex task such as segmenting and labeling of concepts taken from a domain ontology; and evaluation using source language annotation. To test the applicability and effectiveness of the crowdsourced annotation transfer we have considered the case of close and distant language pairs: Italian–Spanish and Italian–Greek. The corpora annotated via crowdsourcing are evaluated against source and target language expert annotations. We demonstrate that the two evaluation references (source and target) highly correlate with each other; thus, drastically reduce the need for the target language reference annotations. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
1574020X
Volume :
52
Issue :
1
Database :
Academic Search Index
Journal :
Language Resources & Evaluation
Publication Type :
Academic Journal
Accession number :
127930801
Full Text :
https://doi.org/10.1007/s10579-017-9396-5