1. Annotator Expertise and Information Quality in Annotation-based Retrieval
- Author
-
Wern Han Lim and Mark James Carman
- Subjects
Bookmarking ,Computer science ,media_common.quotation_subject ,Information quality ,02 engineering and technology ,Popularity ,World Wide Web ,Annotation ,Resource (project management) ,Wisdom of the crowd ,0202 electrical engineering, electronic engineering, information engineering ,Leverage (statistics) ,020201 artificial intelligence & image processing ,Quality (business) ,media_common - Abstract
This paper investigates the annotation-based retrieval (AR) of World Wide Web (WWW) resources that has been annotated by users on Collaborative Tagging (CT) platforms as a form of user-generated content (UGC). Previous approaches have simply weight the WWW resources according to their popularity, in order to leverage on the inherent wisdom of the crowd (WotC). In this paper, we argue that the popularity alone is not a sufficient indicator of quality since (1) some users are better annotators than the others; (2) resource popularity can be easily inflated by malicious users; and (3) high quality but highly specific resources may exhibit lower popularity than more general ones. Thus, we investigate the indicators of information quality for WWW resources, particularly user annotations that can be used to describe them. This research estimates the user expertise of content annotators in order to infer the information quality of their contributions; by exploring the various signals available on social bookmarking platforms such as the temporal information of annotations. The evaluation in retrieval performance on social bookmarking data shows significant improvements with the estimated user expertise and inferred information quality.
- Published
- 2017
- Full Text
- View/download PDF