Detecting driver phone use leveraging car speakers

  • Authors:
  • Jie Yang;Simon Sidhom;Gayathri Chandrasekaran;Tam Vu;Hongbo Liu;Nicolae Cecan;Yingying Chen;Marco Gruteser;Richard P. Martin

  • Affiliations:
  • Stevens Institute of Technology, Hoboken, NJ, USA;Stevens Institute of Technology, Hoboken, NJ, USA;Rutgers University, North Brunswick, NJ, USA;Rutgers University, North Brunswick, NJ, USA;Stevens Institute of Technology, Hoboken, NJ, USA;Rutgers University, North Brunswick, NJ, USA;Stevens Institute of Technology, Hoboken, NJ, USA;Rutgers University, North Brunswick, NJ, USA;Rutgers University, North Brunswick, NJ, USA

  • Venue:
  • MobiCom '11 Proceedings of the 17th annual international conference on Mobile computing and networking
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

This work addresses the fundamental problem of distinguishing between a driver and passenger using a mobile phone, which is the critical input to enable numerous safety and interface enhancements. Our detection system leverages the existing car stereo infrastructure, in particular the speakers and Bluetooth network. Our acoustic approach has the phone send a series of customized high frequency beeps via the car stereo. The beeps are spaced in time across the left, right, and if available, front and rear speakers. After sampling the beeps, we use a sequential change-point detection scheme to time their arrival, and then use a differential approach to estimate the phone's distance from the car's center. From these differences a passenger or driver classification can be made. To validate our approach, we experimented with two kinds of phones and in two different cars. We found that our customized beeps were imperceptible to most users, yet still playable and recordable in both cars. Our customized beeps were also robust to background sounds such as music and wind, and we found the signal processing did not require excessive computational resources. In spite of the cars' heavy multi-path environment, our approach had a classification accuracy of over 90%, and around 95% with some calibrations. We also found we have a low false positive rate, on the order of a few percent.