Human vs machine: establishing a human baseline for multimodal location estimation

  • Authors:
  • Jaeyoung Choi;Howard Lei;Venkatesan Ekambaram;Pascal Kelm;Luke Gottlieb;Thomas Sikora;Kannan Ramchandran;Gerald Friedland

  • Affiliations:
  • International Computer Science Institute, Berkeley, CA, USA;International Computer Science Institute, Berkeley, CA, USA;University of California at Berkeley, Berkeley, CA, USA;Technische Universitat Berlin, Berlin, Germany;International Computer Science Institute, Berkeley, CA, USA;Technische Universitat Berlin, Berlin, Germany;University of California at Berkeley, Berkeley, CA, USA;International Computer Science Institute, Berkeley, CA, USA

  • Venue:
  • Proceedings of the 21st ACM international conference on Multimedia
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Over the recent years, the problem of video location estimation (i.e., estimating the longitude/latitude coordinates of a video without GPS information) has been approached with diverse methods and ideas in the research community and significant improvements have been made. So far, however, systems have only been compared against each other and no systematic study on human performance has been conducted. Based on a human-subject study with 11,900 experiments, this article presents a human baseline for location estimation for different combinations of modalities (audio, audio/video, audio/video/text). Furthermore, this article compares state-of-the-art location estimation systems with the human baseline. Although the overall performance of humans' multimodal video location estimation is better than current machine learning approaches, the difference is quite small: For 41% of the test set, the machine's accuracy was superior to the humans. We present case studies and discuss why machines did better for some videos and not for others. Our analysis suggests new directions and priorities for future work on the improvement of location inference algorithms.