Do you need experts in the crowd?: a case study in image annotation for marine biology

  • Authors:
  • Jiyin He;Jacco van Ossenbruggen;Arjen P. de Vries

  • Affiliations:
  • Centrum Wiskunde en Informatica, Amsterdam, the Netherlands;Centrum Wiskunde en Informatica, Amsterdam, the Netherlands;Centrum Wiskunde en Informatica, Amsterdam, the Netherlands

  • Venue:
  • Proceedings of the 10th Conference on Open Research Areas in Information Retrieval
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Labeled data is a prerequisite for successfully applying machine learning techniques to a wide range of problems. Recently, crowd-sourcing has shown to provide effective solutions to many labeling tasks. However, tasks in specialist domains are difficult to map to Human Intelligence Tasks (or HITs) that can be solved adequately by "the crowd". The question addressed in this paper is whether these specialist tasks can be cast in such a way, that accurate results can still be obtained through crowd-sourcing. We study a case where the goal is to identify fish species in images extracted from videos taken by underwater cameras, a task that typically requires profound domain knowledge in marine biology and hence would be difficult, if not impossible, for the crowd. We show that by carefully converting the recognition task to a visual similarity comparison task, the crowd achieves agreement with the experts comparable to the agreement achieved among experts. Further, non-expert users can learn and improve their performance during the labeling process, e.g., from the system feedback.