Crowd++: unsupervised speaker count with smartphones

  • Authors:
  • Chenren Xu;Sugang Li;Gang Liu;Yanyong Zhang;Emiliano Miluzzo;Yih-Farn Chen;Jun Li;Bernhard Firner

  • Affiliations:
  • Rutgers University, North Brunswick , NJ, USA;Rutgers University, North Brunswick , NJ, USA;UT Dallas, Richardson, NJ, USA;Rutgers University, North Brunswick, NJ, USA;AT&T Labs Research, Florham Park, NJ, USA;AT&T Labs Research, Florham Park, NJ, USA;Rutgers University, North Brunswick, NJ, USA;Rutgers University, North Brunswick, NJ, USA

  • Venue:
  • Proceedings of the 2013 ACM international joint conference on Pervasive and ubiquitous computing
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Smartphones are excellent mobile sensing platforms, with the microphone in particular being exercised in several audio inference applications. We take smartphone audio inference a step further and demonstrate for the first time that it's possible to accurately estimate the number of people talking in a certain place -- with an average error distance of 1.5 speakers -- through unsupervised machine learning analysis on audio segments captured by the smartphones. Inference occurs transparently to the user and no human intervention is needed to derive the classification model. Our results are based on the design, implementation, and evaluation of a system called Crowd++, involving 120 participants in 10 very different environments. We show that no dedicated external hardware or cumbersome supervised learning approaches are needed but only off-the-shelf smartphones used in a transparent manner. We believe our findings have profound implications in many research fields, including social sensing and personal wellbeing assessment.