Haptic and AR interface for paratrooper coordination

  • Authors:
  • Danielle Cummings;George Lucchese;Manoj Prasad;Chris Aikens;Jimmy Ho;Tracy Hammond

  • Affiliations:
  • Texas A&M University, College Station, TX;Texas A&M University, College Station, TX;Texas A&M University, College Station, TX;Texas A&M University, College Station, TX;Texas A&M University, College Station, TX;Texas A&M University, College Station, TX

  • Venue:
  • Proceedings of the 13th International Conference of the NZ Chapter of the ACM's Special Interest Group on Human-Computer Interaction
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Applications that use geolocation data are becoming a common addition to GPS-enabled devices. In terms of mobile computing, there is extensive research in progress to create human-computer interfaces that integrate seamlessly with the user's tasks. When viewing location-based data in a real-world environment, a natural interaction would be to allow the user to see relevant information based on his or her location within an environment. In this paper, we discuss the use of a multi-modal interface that uses haptic feedback and augmented reality to deliver navigation information to paratroopers in the field. This interface was developed for GeoTrooper, a location-based tracking system that visualizes GPS data broadcast by mobile beacons.