首页 | 本学科首页   官方微博 | 高级检索  
     


Limitations of Majority Agreement in Crowdsourced Image Interpretation
Authors:Carl F Salk  Tobias Sturn  Linda See  Steffen Fritz
Affiliation:1. International Institute for Applied Systems Analysis, Laxenburg, Austria;2. Swedish University of Agricultural Sciences, Alnarp, Sweden
Abstract:Crowdsourcing can efficiently complete tasks that are difficult to automate, but the quality of crowdsourced data is tricky to evaluate. Algorithms to grade volunteer work often assume that all tasks are similarly difficult, an assumption that is frequently false. We use a cropland identification game with over 2,600 participants and 165,000 unique tasks to investigate how best to evaluate the difficulty of crowdsourced tasks and to what extent this is possible based on volunteer responses alone. Inter‐volunteer agreement exceeded 90% for about 80% of the images and was negatively correlated with volunteer‐expressed uncertainty about image classification. A total of 343 relatively difficult images were independently classified as cropland, non‐cropland or impossible by two experts. The experts disagreed weakly (one said impossible while the other rated as cropland or non‐cropland) on 27% of the images, but disagreed strongly (cropland vs. non‐cropland) on only 7%. Inter‐volunteer disagreement increased significantly with inter‐expert disagreement. While volunteers agreed with expert classifications for most images, over 20% would have been mis‐categorized if only the volunteers’ majority vote was used. We end with a series of recommendations for managing the challenges posed by heterogeneous tasks in crowdsourcing campaigns.
Keywords:
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号